[ 495.580800] env[62600]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62600) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 495.581152] env[62600]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62600) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 495.581354] env[62600]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62600) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 495.581597] env[62600]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 495.674951] env[62600]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62600) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 495.685248] env[62600]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62600) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 496.288082] env[62600]: INFO nova.virt.driver [None req-ea9c45e6-606a-4b6d-adda-83ce87794e16 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 496.357845] env[62600]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 496.358103] env[62600]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 496.358103] env[62600]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62600) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 499.463809] env[62600]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-71787900-55aa-4132-8f28-809af0b83f31 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.479618] env[62600]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62600) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 499.479797] env[62600]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-3b0df892-6db2-4ead-b848-9adead3adfc4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.520064] env[62600]: INFO oslo_vmware.api [-] Successfully established new session; session ID is cb12c. [ 499.520228] env[62600]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.162s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 499.520727] env[62600]: INFO nova.virt.vmwareapi.driver [None req-ea9c45e6-606a-4b6d-adda-83ce87794e16 None None] VMware vCenter version: 7.0.3 [ 499.524491] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-819ae045-7a77-4aa9-a8c9-458e6a8e6331 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.545486] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-495f51b7-d6a3-47ab-889b-8b4c911d675e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.551278] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60ebfbe1-6096-404a-b58a-76d4bbc4929e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.557655] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80aeef38-bf3f-488f-91c4-4ca0681a96ca {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.570606] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cacd3f1-2dda-4ad0-bb8f-fbc0970918dc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.576289] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55cb63f2-1432-44a8-9249-f30f127516bf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.606357] env[62600]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-c857cc34-6935-405f-9bf2-a23888023774 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.611090] env[62600]: DEBUG nova.virt.vmwareapi.driver [None req-ea9c45e6-606a-4b6d-adda-83ce87794e16 None None] Extension org.openstack.compute already exists. {{(pid=62600) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 499.613750] env[62600]: INFO nova.compute.provider_config [None req-ea9c45e6-606a-4b6d-adda-83ce87794e16 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 500.117243] env[62600]: DEBUG nova.context [None req-ea9c45e6-606a-4b6d-adda-83ce87794e16 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),679522e5-1668-413a-8166-6697c6c0dd81(cell1) {{(pid=62600) load_cells /opt/stack/nova/nova/context.py:464}} [ 500.119411] env[62600]: DEBUG oslo_concurrency.lockutils [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 500.119635] env[62600]: DEBUG oslo_concurrency.lockutils [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 500.127460] env[62600]: DEBUG oslo_concurrency.lockutils [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.008s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 500.127902] env[62600]: DEBUG oslo_concurrency.lockutils [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] Acquiring lock "679522e5-1668-413a-8166-6697c6c0dd81" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 500.128176] env[62600]: DEBUG oslo_concurrency.lockutils [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] Lock "679522e5-1668-413a-8166-6697c6c0dd81" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 500.129207] env[62600]: DEBUG oslo_concurrency.lockutils [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] Lock "679522e5-1668-413a-8166-6697c6c0dd81" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 500.164562] env[62600]: INFO dbcounter [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] Registered counter for database nova_cell0 [ 500.172877] env[62600]: INFO dbcounter [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] Registered counter for database nova_cell1 [ 500.176503] env[62600]: DEBUG oslo_db.sqlalchemy.engines [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62600) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 500.177147] env[62600]: DEBUG oslo_db.sqlalchemy.engines [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62600) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 500.181964] env[62600]: ERROR nova.db.main.api [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 500.181964] env[62600]: result = function(*args, **kwargs) [ 500.181964] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 500.181964] env[62600]: return func(*args, **kwargs) [ 500.181964] env[62600]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 500.181964] env[62600]: result = fn(*args, **kwargs) [ 500.181964] env[62600]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 500.181964] env[62600]: return f(*args, **kwargs) [ 500.181964] env[62600]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 500.181964] env[62600]: return db.service_get_minimum_version(context, binaries) [ 500.181964] env[62600]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 500.181964] env[62600]: _check_db_access() [ 500.181964] env[62600]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 500.181964] env[62600]: stacktrace = ''.join(traceback.format_stack()) [ 500.181964] env[62600]: [ 500.183105] env[62600]: ERROR nova.db.main.api [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 500.183105] env[62600]: result = function(*args, **kwargs) [ 500.183105] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 500.183105] env[62600]: return func(*args, **kwargs) [ 500.183105] env[62600]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 500.183105] env[62600]: result = fn(*args, **kwargs) [ 500.183105] env[62600]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 500.183105] env[62600]: return f(*args, **kwargs) [ 500.183105] env[62600]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 500.183105] env[62600]: return db.service_get_minimum_version(context, binaries) [ 500.183105] env[62600]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 500.183105] env[62600]: _check_db_access() [ 500.183105] env[62600]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 500.183105] env[62600]: stacktrace = ''.join(traceback.format_stack()) [ 500.183105] env[62600]: [ 500.183582] env[62600]: WARNING nova.objects.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 500.183647] env[62600]: WARNING nova.objects.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] Failed to get minimum service version for cell 679522e5-1668-413a-8166-6697c6c0dd81 [ 500.184112] env[62600]: DEBUG oslo_concurrency.lockutils [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] Acquiring lock "singleton_lock" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 500.184282] env[62600]: DEBUG oslo_concurrency.lockutils [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] Acquired lock "singleton_lock" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 500.184528] env[62600]: DEBUG oslo_concurrency.lockutils [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] Releasing lock "singleton_lock" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 500.184853] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] Full set of CONF: {{(pid=62600) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 500.185006] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ******************************************************************************** {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 500.185141] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] Configuration options gathered from: {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 500.185278] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 500.185470] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 500.185601] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ================================================================================ {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 500.185845] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] allow_resize_to_same_host = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.186035] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] arq_binding_timeout = 300 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.186177] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] backdoor_port = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.186307] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] backdoor_socket = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.186474] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] block_device_allocate_retries = 60 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.186647] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] block_device_allocate_retries_interval = 3 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.186840] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cert = self.pem {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.187020] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.187194] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] compute_monitors = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.187367] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] config_dir = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.187539] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] config_drive_format = iso9660 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.187674] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.187839] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] config_source = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.188020] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] console_host = devstack {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.188191] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] control_exchange = nova {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.188350] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cpu_allocation_ratio = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.188510] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] daemon = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.188678] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] debug = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.188838] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] default_access_ip_network_name = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.189011] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] default_availability_zone = nova {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.189176] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] default_ephemeral_format = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.189337] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] default_green_pool_size = 1000 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.189572] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.189738] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] default_schedule_zone = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.189896] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] disk_allocation_ratio = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.190066] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] enable_new_services = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.190248] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] enabled_apis = ['osapi_compute'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.190413] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] enabled_ssl_apis = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.190573] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] flat_injected = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.190732] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] force_config_drive = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.190891] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] force_raw_images = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.191070] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] graceful_shutdown_timeout = 5 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.191237] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] heal_instance_info_cache_interval = 60 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.191451] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] host = cpu-1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.191631] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.191799] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] initial_disk_allocation_ratio = 1.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.191959] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] initial_ram_allocation_ratio = 1.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.192341] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.192527] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] instance_build_timeout = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.192695] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] instance_delete_interval = 300 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.192870] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] instance_format = [instance: %(uuid)s] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.193051] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] instance_name_template = instance-%08x {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.193223] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] instance_usage_audit = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.193400] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] instance_usage_audit_period = month {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.193569] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.193774] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] instances_path = /opt/stack/data/nova/instances {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.193953] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] internal_service_availability_zone = internal {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.194127] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] key = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.194289] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] live_migration_retry_count = 30 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.194460] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] log_color = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.194626] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] log_config_append = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.194799] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.194960] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] log_dir = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.195131] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] log_file = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.195262] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] log_options = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.195425] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] log_rotate_interval = 1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.195593] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] log_rotate_interval_type = days {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.195792] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] log_rotation_type = none {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.195934] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.196077] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.196949] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.197206] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.197356] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.197537] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] long_rpc_timeout = 1800 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.197706] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] max_concurrent_builds = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.197872] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] max_concurrent_live_migrations = 1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.198047] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] max_concurrent_snapshots = 5 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.198215] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] max_local_block_devices = 3 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.198377] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] max_logfile_count = 30 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.198540] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] max_logfile_size_mb = 200 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.198702] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] maximum_instance_delete_attempts = 5 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.198875] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] metadata_listen = 0.0.0.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.199056] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] metadata_listen_port = 8775 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.199231] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] metadata_workers = 2 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.199394] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] migrate_max_retries = -1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.199562] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] mkisofs_cmd = genisoimage {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.199776] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] my_block_storage_ip = 10.180.1.21 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.199914] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] my_ip = 10.180.1.21 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.200093] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] network_allocate_retries = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.200283] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.200461] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] osapi_compute_listen = 0.0.0.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.200626] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] osapi_compute_listen_port = 8774 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.200796] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] osapi_compute_unique_server_name_scope = {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.200968] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] osapi_compute_workers = 2 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.201149] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] password_length = 12 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.201311] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] periodic_enable = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.201472] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] periodic_fuzzy_delay = 60 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.201642] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] pointer_model = usbtablet {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.201814] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] preallocate_images = none {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.201976] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] publish_errors = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.202121] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] pybasedir = /opt/stack/nova {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.202365] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ram_allocation_ratio = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.202543] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] rate_limit_burst = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.202719] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] rate_limit_except_level = CRITICAL {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.202884] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] rate_limit_interval = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.203141] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] reboot_timeout = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.203328] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] reclaim_instance_interval = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.203492] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] record = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.203688] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] reimage_timeout_per_gb = 60 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.203875] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] report_interval = 120 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.204060] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] rescue_timeout = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.204230] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] reserved_host_cpus = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.204392] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] reserved_host_disk_mb = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.204557] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] reserved_host_memory_mb = 512 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.204719] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] reserved_huge_pages = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.204883] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] resize_confirm_window = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.205056] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] resize_fs_using_block_device = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.205223] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] resume_guests_state_on_host_boot = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.205393] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.205557] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] rpc_response_timeout = 60 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.205748] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] run_external_periodic_tasks = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.205937] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] running_deleted_instance_action = reap {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.206115] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] running_deleted_instance_poll_interval = 1800 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.206280] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] running_deleted_instance_timeout = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.206440] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] scheduler_instance_sync_interval = 120 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.206608] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] service_down_time = 720 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.206779] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] servicegroup_driver = db {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.206939] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] shell_completion = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.207112] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] shelved_offload_time = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.207274] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] shelved_poll_interval = 3600 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.207441] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] shutdown_timeout = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.207602] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] source_is_ipv6 = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.207759] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ssl_only = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.208036] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.208216] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] sync_power_state_interval = 600 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.208380] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] sync_power_state_pool_size = 1000 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.208553] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] syslog_log_facility = LOG_USER {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.208726] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] tempdir = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.208873] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] timeout_nbd = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.209054] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] transport_url = **** {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.209221] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] update_resources_interval = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.209382] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] use_cow_images = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.209544] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] use_eventlog = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.209705] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] use_journal = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.209866] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] use_json = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.210036] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] use_rootwrap_daemon = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.210197] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] use_stderr = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.210359] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] use_syslog = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.210518] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vcpu_pin_set = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.210687] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vif_plugging_is_fatal = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.210856] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vif_plugging_timeout = 300 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.211036] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] virt_mkfs = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.211203] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] volume_usage_poll_interval = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.211364] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] watch_log_file = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.211532] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] web = /usr/share/spice-html5 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 500.211723] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.211891] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.212065] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.212241] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_concurrency.disable_process_locking = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.212871] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.213086] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.213354] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.213559] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.213801] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.213987] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.214195] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.auth_strategy = keystone {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.214373] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.compute_link_prefix = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.214554] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.214737] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.dhcp_domain = novalocal {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.214914] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.enable_instance_password = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.215169] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.glance_link_prefix = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.215357] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.215540] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.215733] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.instance_list_per_project_cells = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.215911] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.list_records_by_skipping_down_cells = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.216092] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.local_metadata_per_cell = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.216272] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.max_limit = 1000 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.216444] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.metadata_cache_expiration = 15 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.216622] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.neutron_default_tenant_id = default {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.216799] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.response_validation = warn {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.216978] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.use_neutron_default_nets = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.217168] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.217337] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.217508] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.217686] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.217862] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.vendordata_dynamic_targets = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.218039] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.vendordata_jsonfile_path = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.218228] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.218431] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.backend = dogpile.cache.memcached {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.218605] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.backend_argument = **** {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.218782] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.config_prefix = cache.oslo {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.218959] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.dead_timeout = 60.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.219140] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.debug_cache_backend = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.219307] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.enable_retry_client = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.219473] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.enable_socket_keepalive = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.219637] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.enabled = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.219804] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.enforce_fips_mode = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.219971] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.expiration_time = 600 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.220153] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.hashclient_retry_attempts = 2 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.220321] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.hashclient_retry_delay = 1.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.220488] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.memcache_dead_retry = 300 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.220652] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.memcache_password = **** {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.220820] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.220983] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.221166] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.memcache_pool_maxsize = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.221331] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.221494] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.memcache_sasl_enabled = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.221678] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.221848] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.memcache_socket_timeout = 1.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.222015] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.memcache_username = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.222233] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.proxies = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.222486] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.redis_db = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.222675] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.redis_password = **** {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.222857] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.redis_sentinel_service_name = mymaster {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.223054] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.223296] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.redis_server = localhost:6379 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.223490] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.redis_socket_timeout = 1.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.223657] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.redis_username = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.223829] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.retry_attempts = 2 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.223994] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.retry_delay = 0.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.224174] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.socket_keepalive_count = 1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.224338] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.socket_keepalive_idle = 1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.224500] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.socket_keepalive_interval = 1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.224660] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.tls_allowed_ciphers = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.224822] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.tls_cafile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.224982] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.tls_certfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.225159] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.tls_enabled = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.225321] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cache.tls_keyfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.225494] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cinder.auth_section = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.225697] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cinder.auth_type = password {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.225887] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cinder.cafile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.226079] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cinder.catalog_info = volumev3::publicURL {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.226249] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cinder.certfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.226417] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cinder.collect_timing = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.226579] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cinder.cross_az_attach = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.226744] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cinder.debug = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.226908] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cinder.endpoint_template = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.227087] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cinder.http_retries = 3 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.227255] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cinder.insecure = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.227419] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cinder.keyfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.227593] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cinder.os_region_name = RegionOne {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.227761] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cinder.split_loggers = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.227924] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cinder.timeout = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.228113] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.228279] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] compute.cpu_dedicated_set = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.228441] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] compute.cpu_shared_set = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.228609] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] compute.image_type_exclude_list = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.228773] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.228940] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] compute.max_concurrent_disk_ops = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.229116] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] compute.max_disk_devices_to_attach = -1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.229285] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.229457] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.229620] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] compute.resource_provider_association_refresh = 300 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.229782] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.229946] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] compute.shutdown_retry_interval = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.230424] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.230424] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] conductor.workers = 2 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.230492] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] console.allowed_origins = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.230646] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] console.ssl_ciphers = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.230823] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] console.ssl_minimum_version = default {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.230994] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] consoleauth.enforce_session_timeout = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.231179] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] consoleauth.token_ttl = 600 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.231352] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cyborg.cafile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.231513] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cyborg.certfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.231677] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cyborg.collect_timing = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.231839] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cyborg.connect_retries = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.231998] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cyborg.connect_retry_delay = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.232172] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cyborg.endpoint_override = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.232337] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cyborg.insecure = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.232566] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cyborg.keyfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.232750] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cyborg.max_version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.232920] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cyborg.min_version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.233094] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cyborg.region_name = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.233260] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cyborg.retriable_status_codes = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.233510] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cyborg.service_name = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.233713] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cyborg.service_type = accelerator {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.233892] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cyborg.split_loggers = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.234069] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cyborg.status_code_retries = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.234233] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cyborg.status_code_retry_delay = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.234392] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cyborg.timeout = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.234576] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.234741] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] cyborg.version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.234927] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] database.backend = sqlalchemy {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.235114] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] database.connection = **** {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.235285] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] database.connection_debug = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.235459] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] database.connection_parameters = {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.235633] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] database.connection_recycle_time = 3600 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.235822] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] database.connection_trace = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.235996] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] database.db_inc_retry_interval = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.236177] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] database.db_max_retries = 20 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.236344] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] database.db_max_retry_interval = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.236507] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] database.db_retry_interval = 1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.236709] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] database.max_overflow = 50 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.236922] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] database.max_pool_size = 5 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.237111] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] database.max_retries = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.237292] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.237457] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] database.mysql_wsrep_sync_wait = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.237616] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] database.pool_timeout = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.237783] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] database.retry_interval = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.237949] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] database.slave_connection = **** {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.238124] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] database.sqlite_synchronous = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.238290] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] database.use_db_reconnect = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.238472] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api_database.backend = sqlalchemy {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.238644] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api_database.connection = **** {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.238814] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api_database.connection_debug = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.238985] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api_database.connection_parameters = {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.239164] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api_database.connection_recycle_time = 3600 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.239330] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api_database.connection_trace = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.239491] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api_database.db_inc_retry_interval = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.239655] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api_database.db_max_retries = 20 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.239820] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api_database.db_max_retry_interval = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.239982] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api_database.db_retry_interval = 1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.240159] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api_database.max_overflow = 50 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.240322] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api_database.max_pool_size = 5 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.240484] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api_database.max_retries = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.240656] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.240819] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.240976] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api_database.pool_timeout = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.241152] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api_database.retry_interval = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.241314] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api_database.slave_connection = **** {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.241481] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] api_database.sqlite_synchronous = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.241659] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] devices.enabled_mdev_types = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.241843] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.242020] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ephemeral_storage_encryption.default_format = luks {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.242198] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ephemeral_storage_encryption.enabled = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.242364] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.242535] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.api_servers = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.242789] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.cafile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.242966] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.certfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.243148] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.collect_timing = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.243314] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.connect_retries = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.243552] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.connect_retry_delay = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.243763] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.debug = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.243948] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.default_trusted_certificate_ids = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.244131] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.enable_certificate_validation = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.244302] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.enable_rbd_download = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.244465] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.endpoint_override = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.244633] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.insecure = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.244804] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.keyfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.244963] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.max_version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.245137] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.min_version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.245303] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.num_retries = 3 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.245473] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.rbd_ceph_conf = {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.245643] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.rbd_connect_timeout = 5 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.245837] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.rbd_pool = {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.246026] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.rbd_user = {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.246194] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.region_name = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.246358] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.retriable_status_codes = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.246517] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.service_name = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.246688] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.service_type = image {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.246855] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.split_loggers = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.247029] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.status_code_retries = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.247194] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.status_code_retry_delay = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.247351] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.timeout = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.247532] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.247697] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.verify_glance_signatures = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.247859] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] glance.version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.248082] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] guestfs.debug = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.248317] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] mks.enabled = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.248683] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.248882] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] image_cache.manager_interval = 2400 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.249068] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] image_cache.precache_concurrency = 1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.249245] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] image_cache.remove_unused_base_images = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.249417] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.249587] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.249768] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] image_cache.subdirectory_name = _base {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.249950] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.api_max_retries = 60 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.250132] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.api_retry_interval = 2 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.250296] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.auth_section = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.250461] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.auth_type = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.250625] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.cafile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.250786] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.certfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.250953] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.collect_timing = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.251131] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.conductor_group = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.251292] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.connect_retries = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.251451] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.connect_retry_delay = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.251612] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.endpoint_override = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.251798] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.insecure = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.251934] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.keyfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.252110] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.max_version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.252272] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.min_version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.252440] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.peer_list = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.252601] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.region_name = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.252847] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.retriable_status_codes = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.253042] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.serial_console_state_timeout = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.253214] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.service_name = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.253391] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.service_type = baremetal {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.253637] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.shard = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.253848] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.split_loggers = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.254029] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.status_code_retries = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.254200] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.status_code_retry_delay = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.254363] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.timeout = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.254550] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.254717] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ironic.version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.254908] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.255098] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] key_manager.fixed_key = **** {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.255290] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.255458] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican.barbican_api_version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.255626] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican.barbican_endpoint = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.255833] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican.barbican_endpoint_type = public {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.256020] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican.barbican_region_name = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.256185] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican.cafile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.256347] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican.certfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.256511] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican.collect_timing = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.256699] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican.insecure = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.256881] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican.keyfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.257061] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican.number_of_retries = 60 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.257231] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican.retry_delay = 1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.257398] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican.send_service_user_token = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.257560] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican.split_loggers = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.257721] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican.timeout = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.257885] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican.verify_ssl = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.258052] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican.verify_ssl_path = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.258223] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican_service_user.auth_section = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.258389] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican_service_user.auth_type = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.258547] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican_service_user.cafile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.258706] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican_service_user.certfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.258921] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican_service_user.collect_timing = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.259134] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican_service_user.insecure = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.259302] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican_service_user.keyfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.259470] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican_service_user.split_loggers = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.259629] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] barbican_service_user.timeout = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.259800] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vault.approle_role_id = **** {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.260197] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vault.approle_secret_id = **** {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.260197] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vault.kv_mountpoint = secret {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.260297] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vault.kv_path = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.260674] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vault.kv_version = 2 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.260674] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vault.namespace = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.260794] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vault.root_token_id = **** {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.260923] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vault.ssl_ca_crt_file = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.261113] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vault.timeout = 60.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.261279] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vault.use_ssl = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.261450] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.261637] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.auth_section = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.261782] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.auth_type = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.261944] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.cafile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.262158] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.certfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.262275] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.collect_timing = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.262434] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.connect_retries = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.262594] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.connect_retry_delay = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.262816] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.endpoint_override = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.263016] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.insecure = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.263189] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.keyfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.263351] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.max_version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.263511] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.min_version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.263764] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.region_name = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.263957] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.retriable_status_codes = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.264138] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.service_name = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.264314] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.service_type = identity {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.264479] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.split_loggers = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.264640] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.status_code_retries = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.264803] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.status_code_retry_delay = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.264963] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.timeout = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.265161] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.265325] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] keystone.version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.265526] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.connection_uri = {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.265717] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.cpu_mode = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.265902] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.cpu_model_extra_flags = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.266087] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.cpu_models = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.266266] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.cpu_power_governor_high = performance {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.266439] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.cpu_power_governor_low = powersave {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.266607] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.cpu_power_management = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.266811] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.266986] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.device_detach_attempts = 8 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.267169] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.device_detach_timeout = 20 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.267339] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.disk_cachemodes = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.267502] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.disk_prefix = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.267670] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.enabled_perf_events = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.267836] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.file_backed_memory = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.268015] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.gid_maps = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.268180] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.hw_disk_discard = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.268339] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.hw_machine_type = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.268511] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.images_rbd_ceph_conf = {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.268677] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.268843] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.269017] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.images_rbd_glance_store_name = {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.269193] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.images_rbd_pool = rbd {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.269365] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.images_type = default {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.269526] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.images_volume_group = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.269688] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.inject_key = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.269888] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.inject_partition = -2 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.270085] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.inject_password = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.270267] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.iscsi_iface = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.270420] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.iser_use_multipath = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.270585] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.live_migration_bandwidth = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.270751] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.270918] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.live_migration_downtime = 500 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.271098] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.271260] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.271423] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.live_migration_inbound_addr = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.271587] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.271765] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.live_migration_permit_post_copy = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.271935] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.live_migration_scheme = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.272127] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.live_migration_timeout_action = abort {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.272296] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.live_migration_tunnelled = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.272460] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.live_migration_uri = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.272626] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.live_migration_with_native_tls = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.272787] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.max_queues = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.273046] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.273306] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.273482] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.nfs_mount_options = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.273881] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.274094] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.274273] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.num_iser_scan_tries = 5 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.274442] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.num_memory_encrypted_guests = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.274610] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.274779] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.num_pcie_ports = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.274952] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.num_volume_scan_tries = 5 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.275136] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.pmem_namespaces = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.275304] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.quobyte_client_cfg = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.275593] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.275797] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.rbd_connect_timeout = 5 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.275973] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.276153] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.276319] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.rbd_secret_uuid = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.276481] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.rbd_user = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.276648] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.276827] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.remote_filesystem_transport = ssh {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.276989] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.rescue_image_id = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.277167] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.rescue_kernel_id = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.277332] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.rescue_ramdisk_id = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.277505] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.277668] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.rx_queue_size = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.277839] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.smbfs_mount_options = {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.278138] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.278317] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.snapshot_compression = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.278486] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.snapshot_image_format = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.278713] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.278888] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.sparse_logical_volumes = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.279068] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.swtpm_enabled = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.279256] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.swtpm_group = tss {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.279432] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.swtpm_user = tss {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.279606] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.sysinfo_serial = unique {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.279771] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.tb_cache_size = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.279934] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.tx_queue_size = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.280114] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.uid_maps = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.280280] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.use_virtio_for_bridges = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.280455] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.virt_type = kvm {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.280626] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.volume_clear = zero {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.280788] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.volume_clear_size = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.280956] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.volume_use_multipath = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.281130] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.vzstorage_cache_path = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.281302] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.281472] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.vzstorage_mount_group = qemu {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.281642] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.vzstorage_mount_opts = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.281814] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.282107] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.282293] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.vzstorage_mount_user = stack {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.282463] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.282642] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.auth_section = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.282823] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.auth_type = password {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.283065] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.cafile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.283253] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.certfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.283422] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.collect_timing = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.283586] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.connect_retries = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.283771] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.connect_retry_delay = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.284048] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.default_floating_pool = public {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.284231] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.endpoint_override = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.284402] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.extension_sync_interval = 600 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.284570] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.http_retries = 3 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.284735] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.insecure = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.284902] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.keyfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.285078] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.max_version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.285256] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.285418] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.min_version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.285593] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.ovs_bridge = br-int {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.285795] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.physnets = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.285983] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.region_name = RegionOne {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.286170] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.retriable_status_codes = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.286342] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.service_metadata_proxy = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.286505] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.service_name = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.286681] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.service_type = network {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.286846] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.split_loggers = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.287024] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.status_code_retries = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.287181] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.status_code_retry_delay = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.287341] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.timeout = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.287524] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.287689] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] neutron.version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.287865] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] notifications.bdms_in_notifications = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.288078] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] notifications.default_level = INFO {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.288272] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] notifications.notification_format = unversioned {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.288443] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] notifications.notify_on_state_change = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.288624] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.288804] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] pci.alias = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.288980] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] pci.device_spec = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.289163] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] pci.report_in_placement = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.289335] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.auth_section = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.289509] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.auth_type = password {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.289679] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.289843] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.cafile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.290009] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.certfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.290180] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.collect_timing = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.290341] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.connect_retries = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.290499] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.connect_retry_delay = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.290659] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.default_domain_id = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.290819] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.default_domain_name = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.290978] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.domain_id = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.291149] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.domain_name = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.291310] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.endpoint_override = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.291473] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.insecure = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.291634] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.keyfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.291794] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.max_version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.291951] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.min_version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.292133] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.password = **** {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.292295] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.project_domain_id = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.292462] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.project_domain_name = Default {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.292633] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.project_id = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.292808] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.project_name = service {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.292981] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.region_name = RegionOne {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.293237] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.retriable_status_codes = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.293412] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.service_name = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.293591] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.service_type = placement {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.293791] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.split_loggers = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.294043] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.status_code_retries = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.294236] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.status_code_retry_delay = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.294406] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.system_scope = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.294570] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.timeout = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.294736] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.trust_id = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.294898] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.user_domain_id = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.295081] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.user_domain_name = Default {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.295247] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.user_id = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.295424] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.username = nova {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.295611] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.295810] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] placement.version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.296020] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] quota.cores = 20 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.296192] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] quota.count_usage_from_placement = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.296368] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.296546] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] quota.injected_file_content_bytes = 10240 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.296739] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] quota.injected_file_path_length = 255 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.296918] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] quota.injected_files = 5 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.297101] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] quota.instances = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.297274] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] quota.key_pairs = 100 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.297442] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] quota.metadata_items = 128 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.297607] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] quota.ram = 51200 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.297774] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] quota.recheck_quota = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.297946] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] quota.server_group_members = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.298123] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] quota.server_groups = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.298304] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.298470] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.298638] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] scheduler.image_metadata_prefilter = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.298805] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.298972] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] scheduler.max_attempts = 3 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.299150] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] scheduler.max_placement_results = 1000 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.299319] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.299483] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] scheduler.query_placement_for_image_type_support = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.299649] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.299827] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] scheduler.workers = 2 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.300018] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.300199] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.300391] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.300550] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.300718] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.300888] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.301067] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.301275] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.301454] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.host_subset_size = 1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.301620] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.301782] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.301947] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.302127] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.isolated_hosts = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.302297] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.isolated_images = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.302460] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.302621] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.302782] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.302946] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.pci_in_placement = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.303180] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.303373] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.303541] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.303733] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.303915] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.304164] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.304353] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.track_instance_changes = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.304535] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.304710] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] metrics.required = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.304880] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] metrics.weight_multiplier = 1.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.305061] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.305234] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] metrics.weight_setting = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.305560] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.305766] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] serial_console.enabled = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.305959] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] serial_console.port_range = 10000:20000 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.306150] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.306324] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.306494] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] serial_console.serialproxy_port = 6083 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.306667] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] service_user.auth_section = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.306846] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] service_user.auth_type = password {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.307015] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] service_user.cafile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.307181] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] service_user.certfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.307345] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] service_user.collect_timing = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.307508] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] service_user.insecure = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.307671] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] service_user.keyfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.307843] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] service_user.send_service_user_token = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.308027] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] service_user.split_loggers = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.308202] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] service_user.timeout = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.308374] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] spice.agent_enabled = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.308540] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] spice.enabled = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.308855] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.309066] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.309244] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] spice.html5proxy_port = 6082 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.309412] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] spice.image_compression = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.309575] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] spice.jpeg_compression = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.309738] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] spice.playback_compression = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.309905] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] spice.require_secure = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.310090] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] spice.server_listen = 127.0.0.1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.310268] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.310433] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] spice.streaming_mode = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.310595] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] spice.zlib_compression = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.310764] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] upgrade_levels.baseapi = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.310938] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] upgrade_levels.compute = auto {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.311113] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] upgrade_levels.conductor = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.311275] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] upgrade_levels.scheduler = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.311444] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vendordata_dynamic_auth.auth_section = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.311608] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vendordata_dynamic_auth.auth_type = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.311769] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vendordata_dynamic_auth.cafile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.311932] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vendordata_dynamic_auth.certfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.312107] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.312272] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vendordata_dynamic_auth.insecure = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.312432] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vendordata_dynamic_auth.keyfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.312595] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.312755] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vendordata_dynamic_auth.timeout = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.312929] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.api_retry_count = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.313104] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.ca_file = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.313354] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.cache_prefix = devstack-image-cache {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.313544] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.cluster_name = testcl1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.313736] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.connection_pool_size = 10 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.313911] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.console_delay_seconds = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.314098] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.datastore_regex = ^datastore.* {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.314402] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.314589] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.host_password = **** {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.314763] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.host_port = 443 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.314938] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.host_username = administrator@vsphere.local {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.315124] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.insecure = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.315291] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.integration_bridge = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.315459] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.maximum_objects = 100 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.315620] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.pbm_default_policy = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.315926] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.pbm_enabled = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.316077] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.pbm_wsdl_location = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.316258] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.316420] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.serial_port_proxy_uri = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.316581] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.serial_port_service_uri = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.316755] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.task_poll_interval = 0.5 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.316930] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.use_linked_clone = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.317119] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.vnc_keymap = en-us {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.317294] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.vnc_port = 5900 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.317461] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vmware.vnc_port_total = 10000 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.317652] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vnc.auth_schemes = ['none'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.317832] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vnc.enabled = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.318171] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.318371] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.318549] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vnc.novncproxy_port = 6080 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.318731] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vnc.server_listen = 127.0.0.1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.318910] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.319087] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vnc.vencrypt_ca_certs = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.319253] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vnc.vencrypt_client_cert = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.319414] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vnc.vencrypt_client_key = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.319597] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.319763] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.disable_deep_image_inspection = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.319927] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.320103] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.320269] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.320432] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.disable_rootwrap = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.320594] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.enable_numa_live_migration = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.320756] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.320920] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.321092] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.321259] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.libvirt_disable_apic = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.321422] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.321584] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.321747] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.321913] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.322087] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.322250] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.322413] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.322574] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.322734] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.322904] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.323100] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.323274] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] wsgi.client_socket_timeout = 900 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.323529] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] wsgi.default_pool_size = 1000 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.323729] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] wsgi.keep_alive = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.323914] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] wsgi.max_header_line = 16384 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.324129] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] wsgi.secure_proxy_ssl_header = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.324405] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] wsgi.ssl_ca_file = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.324593] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] wsgi.ssl_cert_file = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.324763] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] wsgi.ssl_key_file = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.324938] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] wsgi.tcp_keepidle = 600 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.325144] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.325324] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] zvm.ca_file = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.325486] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] zvm.cloud_connector_url = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.325809] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.326012] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] zvm.reachable_timeout = 300 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.326208] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_policy.enforce_new_defaults = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.326606] env[62600]: WARNING oslo_config.cfg [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 500.326797] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_policy.enforce_scope = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.326981] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_policy.policy_default_rule = default {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.327180] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.327360] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_policy.policy_file = policy.yaml {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.327541] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.327708] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.327873] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.328072] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.328257] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.328431] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.328611] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.328789] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] profiler.connection_string = messaging:// {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.328959] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] profiler.enabled = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.329147] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] profiler.es_doc_type = notification {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.329315] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] profiler.es_scroll_size = 10000 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.329487] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] profiler.es_scroll_time = 2m {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.329652] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] profiler.filter_error_trace = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.329825] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] profiler.hmac_keys = **** {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.329997] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] profiler.sentinel_service_name = mymaster {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.330181] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] profiler.socket_timeout = 0.1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.330345] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] profiler.trace_requests = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.330510] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] profiler.trace_sqlalchemy = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.330693] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] profiler_jaeger.process_tags = {} {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.330856] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] profiler_jaeger.service_name_prefix = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.331029] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] profiler_otlp.service_name_prefix = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.331204] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] remote_debug.host = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.331364] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] remote_debug.port = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.331542] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.331705] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.331871] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.332043] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.332210] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.332371] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.332532] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.332694] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.332858] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.333039] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.333207] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.333383] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.333649] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.333848] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.334040] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.336650] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.336869] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.337079] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.337260] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.337432] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.337606] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.337779] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.337953] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.338139] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.338307] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.338473] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.338639] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.338807] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.338976] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.339160] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.ssl = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.339339] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.339515] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.339682] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.339862] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.340049] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.ssl_version = {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.340222] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.340413] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.340587] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_notifications.retry = -1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.340776] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.340960] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_messaging_notifications.transport_url = **** {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.341152] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.auth_section = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.341318] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.auth_type = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.341480] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.cafile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.341642] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.certfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.341807] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.collect_timing = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.341968] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.connect_retries = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.342146] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.connect_retry_delay = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.342309] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.endpoint_id = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.342469] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.endpoint_override = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.342631] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.insecure = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.342791] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.keyfile = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.342950] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.max_version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.343119] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.min_version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.343281] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.region_name = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.343463] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.retriable_status_codes = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.343711] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.service_name = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.343895] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.service_type = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.344076] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.split_loggers = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.344240] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.status_code_retries = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.344400] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.status_code_retry_delay = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.344558] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.timeout = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.344720] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.valid_interfaces = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.344882] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_limit.version = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.345064] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_reports.file_event_handler = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.345236] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.345400] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] oslo_reports.log_dir = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.345572] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.345767] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.345933] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.346121] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.346292] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.346453] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.346665] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.346832] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vif_plug_ovs_privileged.group = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.346995] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.347175] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.347342] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.347504] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] vif_plug_ovs_privileged.user = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.347676] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] os_vif_linux_bridge.flat_interface = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.347861] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.348050] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.348231] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.348407] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.348577] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.348747] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.348914] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.349106] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.349283] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] os_vif_ovs.isolate_vif = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.349453] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.349621] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.349793] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.349968] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] os_vif_ovs.ovsdb_interface = native {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.350148] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] os_vif_ovs.per_port_bridge = False {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.350320] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] privsep_osbrick.capabilities = [21] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.350484] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] privsep_osbrick.group = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.350645] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] privsep_osbrick.helper_command = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.350812] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.350977] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.351251] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] privsep_osbrick.user = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.351369] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.351538] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] nova_sys_admin.group = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.351702] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] nova_sys_admin.helper_command = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.351869] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.352046] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.352214] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] nova_sys_admin.user = None {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 500.352341] env[62600]: DEBUG oslo_service.service [None req-39cba24d-1548-4c52-91f6-e47e00c42ccc None None] ******************************************************************************** {{(pid=62600) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 500.352841] env[62600]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 500.856789] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Getting list of instances from cluster (obj){ [ 500.856789] env[62600]: value = "domain-c8" [ 500.856789] env[62600]: _type = "ClusterComputeResource" [ 500.856789] env[62600]: } {{(pid=62600) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 500.857993] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00992467-9e2a-47f0-a9b9-e2e777a798db {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.867138] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Got total of 0 instances {{(pid=62600) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 500.867678] env[62600]: WARNING nova.virt.vmwareapi.driver [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 500.868172] env[62600]: INFO nova.virt.node [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Generated node identity 664af347-7147-4bf5-9019-9ae15cb4aa82 [ 500.868407] env[62600]: INFO nova.virt.node [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Wrote node identity 664af347-7147-4bf5-9019-9ae15cb4aa82 to /opt/stack/data/n-cpu-1/compute_id [ 501.371709] env[62600]: WARNING nova.compute.manager [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Compute nodes ['664af347-7147-4bf5-9019-9ae15cb4aa82'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 502.378221] env[62600]: INFO nova.compute.manager [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 503.383595] env[62600]: WARNING nova.compute.manager [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 503.384024] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 503.384119] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 503.384277] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 503.384436] env[62600]: DEBUG nova.compute.resource_tracker [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62600) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 503.385374] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de4950f6-a008-4b04-ab81-1f176a9db2dd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.393783] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ec56302-3916-43cf-866a-5f31f3a9ce57 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.407203] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6bb68a0-51fb-496a-85e7-c266fdb2ecaa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.413446] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00b60041-03fe-4412-acc5-0be30911f0a0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.442623] env[62600]: DEBUG nova.compute.resource_tracker [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181500MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=62600) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 503.442792] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 503.442971] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 503.946151] env[62600]: WARNING nova.compute.resource_tracker [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] No compute node record for cpu-1:664af347-7147-4bf5-9019-9ae15cb4aa82: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 664af347-7147-4bf5-9019-9ae15cb4aa82 could not be found. [ 504.450699] env[62600]: INFO nova.compute.resource_tracker [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 664af347-7147-4bf5-9019-9ae15cb4aa82 [ 505.963181] env[62600]: DEBUG nova.compute.resource_tracker [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 505.963617] env[62600]: DEBUG nova.compute.resource_tracker [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 506.119082] env[62600]: INFO nova.scheduler.client.report [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] [req-008661b6-4bd7-4054-8e4f-97665b16bdaa] Created resource provider record via placement API for resource provider with UUID 664af347-7147-4bf5-9019-9ae15cb4aa82 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 506.136505] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-775f1070-6add-4252-b3ef-ba0348076aff {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.144700] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a53647da-a7a0-44b8-ba61-7043257082be {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.174693] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bd5a26d-cd42-4b4f-a4e1-7adafab1466d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.181551] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d54df347-d436-4d2a-bee5-e4c43081dc02 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.194108] env[62600]: DEBUG nova.compute.provider_tree [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 506.727730] env[62600]: DEBUG nova.scheduler.client.report [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 506.727966] env[62600]: DEBUG nova.compute.provider_tree [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 0 to 1 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 506.728121] env[62600]: DEBUG nova.compute.provider_tree [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 506.777761] env[62600]: DEBUG nova.compute.provider_tree [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 1 to 2 during operation: update_traits {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 507.282801] env[62600]: DEBUG nova.compute.resource_tracker [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62600) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 507.283181] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.840s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 507.283181] env[62600]: DEBUG nova.service [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Creating RPC server for service compute {{(pid=62600) start /opt/stack/nova/nova/service.py:186}} [ 507.297135] env[62600]: DEBUG nova.service [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] Join ServiceGroup membership for this service compute {{(pid=62600) start /opt/stack/nova/nova/service.py:203}} [ 507.297341] env[62600]: DEBUG nova.servicegroup.drivers.db [None req-a9c76a39-b776-4a5b-8abd-afcf323525b8 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62600) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 509.299668] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._sync_power_states {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 509.803439] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Getting list of instances from cluster (obj){ [ 509.803439] env[62600]: value = "domain-c8" [ 509.803439] env[62600]: _type = "ClusterComputeResource" [ 509.803439] env[62600]: } {{(pid=62600) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 509.804657] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b6f0fb-0ea1-4122-806d-66b9a08f6179 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.813570] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Got total of 0 instances {{(pid=62600) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 509.813798] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 509.814139] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Getting list of instances from cluster (obj){ [ 509.814139] env[62600]: value = "domain-c8" [ 509.814139] env[62600]: _type = "ClusterComputeResource" [ 509.814139] env[62600]: } {{(pid=62600) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 509.815114] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e4f03ec-ee66-4083-acb9-959d616d62eb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.822437] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Got total of 0 instances {{(pid=62600) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 546.038829] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Acquiring lock "98cbcc28-54e0-40de-9feb-6da917806423" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.039145] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Lock "98cbcc28-54e0-40de-9feb-6da917806423" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.220525] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Acquiring lock "da07d0e2-98bc-451c-955e-db5eaeee3c82" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.220765] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Lock "da07d0e2-98bc-451c-955e-db5eaeee3c82" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.487521] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Acquiring lock "b315463b-f8aa-4b16-a5a2-069458f15358" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.487521] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Lock "b315463b-f8aa-4b16-a5a2-069458f15358" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.551403] env[62600]: DEBUG nova.compute.manager [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 546.722888] env[62600]: DEBUG nova.compute.manager [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 546.797043] env[62600]: DEBUG oslo_concurrency.lockutils [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Acquiring lock "371a3210-1741-43ce-b635-b84314aee77e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.797353] env[62600]: DEBUG oslo_concurrency.lockutils [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Lock "371a3210-1741-43ce-b635-b84314aee77e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.996644] env[62600]: DEBUG nova.compute.manager [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 547.100717] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.100717] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.105305] env[62600]: INFO nova.compute.claims [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 547.257063] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.300352] env[62600]: DEBUG nova.compute.manager [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 547.310794] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Acquiring lock "12d03944-78d1-44e4-aaf9-5678f7f3e1ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.311083] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Lock "12d03944-78d1-44e4-aaf9-5678f7f3e1ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.511590] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Acquiring lock "5ebc4807-907b-4254-8496-92f1bbb7b761" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.511590] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Lock "5ebc4807-907b-4254-8496-92f1bbb7b761" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.524647] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.819999] env[62600]: DEBUG nova.compute.manager [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 547.827990] env[62600]: DEBUG oslo_concurrency.lockutils [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.013995] env[62600]: DEBUG nova.compute.manager [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 548.255663] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c640fe73-f353-452a-aa1a-a1fde87a9fb1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.263637] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1626acac-1c2d-4fe9-ac75-71bc3743e348 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.299965] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ebf1cc7-ca2b-4b35-a55a-9b9634fbe300 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.311018] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e66ddf-1143-420a-a754-6d722f4de61b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.329686] env[62600]: DEBUG nova.compute.provider_tree [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 548.347801] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.369101] env[62600]: DEBUG oslo_concurrency.lockutils [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "b74bf924-2d0b-4221-9c82-aa5b1ef903ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.369324] env[62600]: DEBUG oslo_concurrency.lockutils [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "b74bf924-2d0b-4221-9c82-aa5b1ef903ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.549495] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.603578] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "a664416f-f9f6-4cdd-93b0-8dd63181562e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.603838] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "a664416f-f9f6-4cdd-93b0-8dd63181562e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.835255] env[62600]: DEBUG nova.scheduler.client.report [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 548.872219] env[62600]: DEBUG nova.compute.manager [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 549.107137] env[62600]: DEBUG nova.compute.manager [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 549.185835] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Acquiring lock "38e57696-33a8-49d8-b061-ae14155ed280" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.186046] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Lock "38e57696-33a8-49d8-b061-ae14155ed280" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.346529] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.246s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.347346] env[62600]: DEBUG nova.compute.manager [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 549.351830] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.096s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.353409] env[62600]: INFO nova.compute.claims [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 549.416414] env[62600]: DEBUG oslo_concurrency.lockutils [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.521803] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Acquiring lock "c6cf703e-5b00-44b3-ae8b-b655105a073b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.521803] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Lock "c6cf703e-5b00-44b3-ae8b-b655105a073b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.640712] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.689119] env[62600]: DEBUG nova.compute.manager [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 549.861907] env[62600]: DEBUG nova.compute.utils [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 549.863915] env[62600]: DEBUG nova.compute.manager [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Not allocating networking since 'none' was specified. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 549.876771] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.877199] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.001912] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Acquiring lock "ec99a26e-ee0a-44da-80b6-afd1205c7ea0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.002172] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Lock "ec99a26e-ee0a-44da-80b6-afd1205c7ea0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.026208] env[62600]: DEBUG nova.compute.manager [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 550.210568] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.366371] env[62600]: DEBUG nova.compute.manager [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 550.438839] env[62600]: DEBUG oslo_concurrency.lockutils [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Acquiring lock "8c0e3a15-0cdd-422a-8a97-95d57621de88" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.439624] env[62600]: DEBUG oslo_concurrency.lockutils [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Lock "8c0e3a15-0cdd-422a-8a97-95d57621de88" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.562131] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.745106] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d923593-68ea-43ec-85f5-1a3d6b66f7b2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.754343] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdfa4cc5-d815-4252-aac7-4a7398108839 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.785825] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9091a7b7-ad46-4ae3-bb53-671a87d728dd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.796474] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f81e3a5e-9162-4114-ac35-df7e4c8d4c4d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.811827] env[62600]: DEBUG nova.compute.provider_tree [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 550.869803] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Acquiring lock "a934e86f-1a51-468e-915f-add473fc2bf6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.870048] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Lock "a934e86f-1a51-468e-915f-add473fc2bf6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.315272] env[62600]: DEBUG nova.scheduler.client.report [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 551.378079] env[62600]: DEBUG nova.compute.manager [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 551.658842] env[62600]: DEBUG nova.virt.hardware [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 551.659066] env[62600]: DEBUG nova.virt.hardware [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 551.659291] env[62600]: DEBUG nova.virt.hardware [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 551.659658] env[62600]: DEBUG nova.virt.hardware [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 551.659658] env[62600]: DEBUG nova.virt.hardware [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 551.659658] env[62600]: DEBUG nova.virt.hardware [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 551.660534] env[62600]: DEBUG nova.virt.hardware [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 551.660534] env[62600]: DEBUG nova.virt.hardware [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 551.661059] env[62600]: DEBUG nova.virt.hardware [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 551.661246] env[62600]: DEBUG nova.virt.hardware [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 551.661431] env[62600]: DEBUG nova.virt.hardware [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 551.662310] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e34535f-5cf0-43aa-902d-f91a446fa0f5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.674168] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01129666-b7fa-4979-a859-6e1e5a8207a8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.696554] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd79dfe7-3141-4c52-b307-be6c5780bca6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.715978] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Instance VIF info [] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 551.725382] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 551.726643] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b4261f58-6064-4e8f-868d-592aada742c0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.737058] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Created folder: OpenStack in parent group-v4. [ 551.737179] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Creating folder: Project (eb8416b9f3ad445a8870db6c54973575). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 551.737442] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f6ea87ba-4b34-4b62-b10d-02b3d8b2ec13 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.747468] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Created folder: Project (eb8416b9f3ad445a8870db6c54973575) in parent group-v264198. [ 551.747701] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Creating folder: Instances. Parent ref: group-v264199. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 551.747927] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4f4f66ff-57fc-4480-9fc9-e60975163f94 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.757996] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Created folder: Instances in parent group-v264199. [ 551.758323] env[62600]: DEBUG oslo.service.loopingcall [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 551.758518] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 551.758717] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b95fbb8c-9161-4eea-9ae7-6a771d7258f4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.778641] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 551.778641] env[62600]: value = "task-1222179" [ 551.778641] env[62600]: _type = "Task" [ 551.778641] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.794368] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222179, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.820480] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.469s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.821057] env[62600]: DEBUG nova.compute.manager [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 551.823634] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.299s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.825085] env[62600]: INFO nova.compute.claims [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 552.105905] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Acquiring lock "b3cc82aa-be39-4e89-8972-b5596629961c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.106203] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Lock "b3cc82aa-be39-4e89-8972-b5596629961c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.291623] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222179, 'name': CreateVM_Task, 'duration_secs': 0.410667} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.291803] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 552.292987] env[62600]: DEBUG oslo_vmware.service [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45364e5b-8562-4171-8978-7249abd5f777 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.299387] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.299387] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.299990] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 552.300272] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55c98c23-532c-48d0-b775-13357838a7bb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.306705] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 552.306705] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ead1d2-9bfe-07c3-ab0d-256c08d9a62b" [ 552.306705] env[62600]: _type = "Task" [ 552.306705] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.314228] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ead1d2-9bfe-07c3-ab0d-256c08d9a62b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.336092] env[62600]: DEBUG nova.compute.utils [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 552.340276] env[62600]: DEBUG nova.compute.manager [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 552.340276] env[62600]: DEBUG nova.network.neutron [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 552.443484] env[62600]: DEBUG nova.policy [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5328c9c7e09f4bb7b1275d9e99329fbd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c754d7b61e7e4225ac177fda4dbbb1e2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 552.818982] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.820921] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 552.820921] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.820921] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.821643] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 552.821643] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0cf0b912-59f7-4dc6-97c7-c81db34ac5fd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.850063] env[62600]: DEBUG nova.compute.manager [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 552.856697] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 552.857043] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 552.857943] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a5fc0cc-0173-481f-8d3a-2fec8b41c998 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.868610] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-463287f2-b322-4aa3-ac95-42762e198f0b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.880079] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 552.880079] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525b184a-023b-bc3b-7294-6bbc41ce2ed8" [ 552.880079] env[62600]: _type = "Task" [ 552.880079] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.894530] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525b184a-023b-bc3b-7294-6bbc41ce2ed8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.127213] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1409908a-2dc9-4afc-a50f-15ce9ae92fef {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.144587] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca097d0-7a54-4142-af98-e0b72a36ad7c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.153265] env[62600]: DEBUG nova.network.neutron [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Successfully created port: 1339efad-11b7-403a-9a9c-9f70795493ea {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 553.184198] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f37fdd2-2178-40ac-b234-b509a539f2a6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.194876] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e45d674-68a7-4768-a874-9b7eae9384ad {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.210480] env[62600]: DEBUG nova.compute.provider_tree [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 553.393889] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Preparing fetch location {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 553.394513] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Creating directory with path [datastore2] vmware_temp/33663aab-fa83-43a8-8e4f-928d936d78bf/e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 553.394513] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-699a913d-8f2e-4518-a238-50392311c5f4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.419519] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Created directory with path [datastore2] vmware_temp/33663aab-fa83-43a8-8e4f-928d936d78bf/e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 553.419684] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Fetch image to [datastore2] vmware_temp/33663aab-fa83-43a8-8e4f-928d936d78bf/e7bd1d77-bfff-4684-9545-f6a1c69efe58/tmp-sparse.vmdk {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 553.419850] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Downloading image file data e7bd1d77-bfff-4684-9545-f6a1c69efe58 to [datastore2] vmware_temp/33663aab-fa83-43a8-8e4f-928d936d78bf/e7bd1d77-bfff-4684-9545-f6a1c69efe58/tmp-sparse.vmdk on the data store datastore2 {{(pid=62600) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 553.420671] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197729e7-9cd3-4d75-9362-07006680d4b0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.429219] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24924390-78f3-453b-a9ee-05317323375d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.440107] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-784e35bc-a442-47ba-8de3-f931225cb444 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.479897] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df71279-64d7-4ce2-9fdd-2db4e6725098 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.490602] env[62600]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-c5e9966b-9475-449d-a473-0880b09dcc0e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.589657] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Downloading image file data e7bd1d77-bfff-4684-9545-f6a1c69efe58 to the data store datastore2 {{(pid=62600) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 553.658232] env[62600]: DEBUG oslo_vmware.rw_handles [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/33663aab-fa83-43a8-8e4f-928d936d78bf/e7bd1d77-bfff-4684-9545-f6a1c69efe58/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62600) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 553.720150] env[62600]: DEBUG nova.scheduler.client.report [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 553.870022] env[62600]: DEBUG nova.compute.manager [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 553.906458] env[62600]: DEBUG nova.virt.hardware [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 553.906701] env[62600]: DEBUG nova.virt.hardware [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 553.906855] env[62600]: DEBUG nova.virt.hardware [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 553.907063] env[62600]: DEBUG nova.virt.hardware [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 553.907330] env[62600]: DEBUG nova.virt.hardware [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 553.907530] env[62600]: DEBUG nova.virt.hardware [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 553.907761] env[62600]: DEBUG nova.virt.hardware [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 553.907940] env[62600]: DEBUG nova.virt.hardware [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 553.908151] env[62600]: DEBUG nova.virt.hardware [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 553.908363] env[62600]: DEBUG nova.virt.hardware [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 553.908512] env[62600]: DEBUG nova.virt.hardware [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 553.909712] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d514be-f3ed-4f17-805c-78fecf1817a0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.925377] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8c7b533-68c7-404d-8e93-d00aecf4d7fb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.227938] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.403s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.228369] env[62600]: DEBUG nova.compute.manager [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 554.236311] env[62600]: DEBUG oslo_concurrency.lockutils [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.404s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.236311] env[62600]: INFO nova.compute.claims [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 554.434080] env[62600]: DEBUG oslo_vmware.rw_handles [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Completed reading data from the image iterator. {{(pid=62600) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 554.434080] env[62600]: DEBUG oslo_vmware.rw_handles [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/33663aab-fa83-43a8-8e4f-928d936d78bf/e7bd1d77-bfff-4684-9545-f6a1c69efe58/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62600) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 554.495082] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Downloaded image file data e7bd1d77-bfff-4684-9545-f6a1c69efe58 to vmware_temp/33663aab-fa83-43a8-8e4f-928d936d78bf/e7bd1d77-bfff-4684-9545-f6a1c69efe58/tmp-sparse.vmdk on the data store datastore2 {{(pid=62600) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 554.497246] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Caching image {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 554.497349] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Copying Virtual Disk [datastore2] vmware_temp/33663aab-fa83-43a8-8e4f-928d936d78bf/e7bd1d77-bfff-4684-9545-f6a1c69efe58/tmp-sparse.vmdk to [datastore2] vmware_temp/33663aab-fa83-43a8-8e4f-928d936d78bf/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 554.498057] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2751f18e-4605-4267-8422-dc1693f21b17 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.509326] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 554.509326] env[62600]: value = "task-1222180" [ 554.509326] env[62600]: _type = "Task" [ 554.509326] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.518578] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222180, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.687632] env[62600]: DEBUG oslo_concurrency.lockutils [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Acquiring lock "ba442811-2b0a-44e7-b036-572291aff648" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.688185] env[62600]: DEBUG oslo_concurrency.lockutils [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Lock "ba442811-2b0a-44e7-b036-572291aff648" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.740939] env[62600]: DEBUG nova.compute.utils [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 554.748031] env[62600]: DEBUG nova.compute.manager [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 554.748244] env[62600]: DEBUG nova.network.neutron [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 554.917755] env[62600]: DEBUG nova.policy [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '99e3acf8403b4db19c846f926c365d7d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3fba72f88d0c430297d71f4b1914b1be', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 555.021681] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222180, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.248686] env[62600]: DEBUG nova.compute.manager [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 555.520712] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222180, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.718089} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.523382] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Copied Virtual Disk [datastore2] vmware_temp/33663aab-fa83-43a8-8e4f-928d936d78bf/e7bd1d77-bfff-4684-9545-f6a1c69efe58/tmp-sparse.vmdk to [datastore2] vmware_temp/33663aab-fa83-43a8-8e4f-928d936d78bf/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 555.523690] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Deleting the datastore file [datastore2] vmware_temp/33663aab-fa83-43a8-8e4f-928d936d78bf/e7bd1d77-bfff-4684-9545-f6a1c69efe58/tmp-sparse.vmdk {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 555.524689] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7fe46a3d-d4c1-4fb6-a7b2-a0853c3d8199 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.533168] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 555.533168] env[62600]: value = "task-1222181" [ 555.533168] env[62600]: _type = "Task" [ 555.533168] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.544610] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222181, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.566168] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f85b59-32f5-460b-96c9-ed9a029fb563 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.573437] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdd795c4-b3a5-4de3-8c17-52fb73881655 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.607385] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ec069e2-a632-4190-a242-bdb66bb0ecd8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.615737] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e367025a-6e6e-4b73-bf75-b1630ade8eb4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.632326] env[62600]: DEBUG nova.compute.provider_tree [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 555.736551] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 555.736551] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 555.736713] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Starting heal instance info cache {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 555.737305] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Rebuilding the list of instances to heal {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 555.744899] env[62600]: DEBUG nova.network.neutron [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Successfully created port: 92c2ae25-d1b4-47d0-a419-86065fc4d35c {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 556.043289] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222181, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024928} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.043548] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 556.043748] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Moving file from [datastore2] vmware_temp/33663aab-fa83-43a8-8e4f-928d936d78bf/e7bd1d77-bfff-4684-9545-f6a1c69efe58 to [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58. {{(pid=62600) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 556.044201] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-255dfe1b-0ff5-4044-baff-a3ca119c79df {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.050866] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 556.050866] env[62600]: value = "task-1222182" [ 556.050866] env[62600]: _type = "Task" [ 556.050866] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.059643] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222182, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.135676] env[62600]: DEBUG nova.scheduler.client.report [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 556.241651] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Skipping network cache update for instance because it is Building. {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 556.241848] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Skipping network cache update for instance because it is Building. {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 556.241981] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Skipping network cache update for instance because it is Building. {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 556.242122] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Skipping network cache update for instance because it is Building. {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 556.242296] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Didn't find any instances for network info cache update. {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 556.242520] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 556.242776] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 556.242945] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 556.243173] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 556.243365] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 556.243552] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 556.243748] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62600) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 556.243915] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 556.263439] env[62600]: DEBUG nova.compute.manager [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 556.289899] env[62600]: DEBUG nova.virt.hardware [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 556.290218] env[62600]: DEBUG nova.virt.hardware [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 556.290382] env[62600]: DEBUG nova.virt.hardware [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 556.290568] env[62600]: DEBUG nova.virt.hardware [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 556.290716] env[62600]: DEBUG nova.virt.hardware [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 556.290861] env[62600]: DEBUG nova.virt.hardware [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 556.291227] env[62600]: DEBUG nova.virt.hardware [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 556.291461] env[62600]: DEBUG nova.virt.hardware [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 556.291649] env[62600]: DEBUG nova.virt.hardware [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 556.291839] env[62600]: DEBUG nova.virt.hardware [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 556.292088] env[62600]: DEBUG nova.virt.hardware [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 556.293017] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc3d2460-1370-4281-9d78-cdcad8781ebb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.302654] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-185edbf9-dea9-4263-8c0f-33d56a9ac745 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.570249] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222182, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.0519} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.570583] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] File moved {{(pid=62600) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 556.570781] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Cleaning up location [datastore2] vmware_temp/33663aab-fa83-43a8-8e4f-928d936d78bf {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 556.570963] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Deleting the datastore file [datastore2] vmware_temp/33663aab-fa83-43a8-8e4f-928d936d78bf {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 556.571346] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3afccc85-3d41-4253-b0c5-cb464e542b35 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.579041] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 556.579041] env[62600]: value = "task-1222183" [ 556.579041] env[62600]: _type = "Task" [ 556.579041] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.588557] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222183, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.644023] env[62600]: ERROR nova.compute.manager [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1339efad-11b7-403a-9a9c-9f70795493ea, please check neutron logs for more information. [ 556.644023] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 556.644023] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.644023] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 556.644023] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 556.644023] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 556.644023] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 556.644023] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 556.644023] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.644023] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 556.644023] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.644023] env[62600]: ERROR nova.compute.manager raise self.value [ 556.644023] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 556.644023] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 556.644023] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.644023] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 556.644520] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.644520] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 556.644520] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1339efad-11b7-403a-9a9c-9f70795493ea, please check neutron logs for more information. [ 556.644520] env[62600]: ERROR nova.compute.manager [ 556.644520] env[62600]: Traceback (most recent call last): [ 556.644520] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 556.644520] env[62600]: listener.cb(fileno) [ 556.644520] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.644520] env[62600]: result = function(*args, **kwargs) [ 556.644520] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.644520] env[62600]: return func(*args, **kwargs) [ 556.644520] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.644520] env[62600]: raise e [ 556.644520] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.644520] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 556.644520] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 556.644520] env[62600]: created_port_ids = self._update_ports_for_instance( [ 556.644520] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 556.644520] env[62600]: with excutils.save_and_reraise_exception(): [ 556.644520] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.644520] env[62600]: self.force_reraise() [ 556.644520] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.644520] env[62600]: raise self.value [ 556.644520] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 556.644520] env[62600]: updated_port = self._update_port( [ 556.644520] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.644520] env[62600]: _ensure_no_port_binding_failure(port) [ 556.644520] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.644520] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 556.645244] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 1339efad-11b7-403a-9a9c-9f70795493ea, please check neutron logs for more information. [ 556.645244] env[62600]: Removing descriptor: 16 [ 556.645367] env[62600]: DEBUG oslo_concurrency.lockutils [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.413s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.645926] env[62600]: DEBUG nova.compute.manager [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 556.650785] env[62600]: ERROR nova.compute.manager [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1339efad-11b7-403a-9a9c-9f70795493ea, please check neutron logs for more information. [ 556.650785] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Traceback (most recent call last): [ 556.650785] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 556.650785] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] yield resources [ 556.650785] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 556.650785] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] self.driver.spawn(context, instance, image_meta, [ 556.650785] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 556.650785] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] self._vmops.spawn(context, instance, image_meta, injected_files, [ 556.650785] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 556.650785] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] vm_ref = self.build_virtual_machine(instance, [ 556.650785] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 556.651725] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] vif_infos = vmwarevif.get_vif_info(self._session, [ 556.651725] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 556.651725] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] for vif in network_info: [ 556.651725] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 556.651725] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] return self._sync_wrapper(fn, *args, **kwargs) [ 556.651725] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 556.651725] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] self.wait() [ 556.651725] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 556.651725] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] self[:] = self._gt.wait() [ 556.651725] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 556.651725] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] return self._exit_event.wait() [ 556.651725] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 556.651725] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] result = hub.switch() [ 556.652710] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 556.652710] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] return self.greenlet.switch() [ 556.652710] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.652710] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] result = function(*args, **kwargs) [ 556.652710] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.652710] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] return func(*args, **kwargs) [ 556.652710] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.652710] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] raise e [ 556.652710] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.652710] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] nwinfo = self.network_api.allocate_for_instance( [ 556.652710] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 556.652710] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] created_port_ids = self._update_ports_for_instance( [ 556.652710] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 556.655189] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] with excutils.save_and_reraise_exception(): [ 556.655189] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.655189] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] self.force_reraise() [ 556.655189] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.655189] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] raise self.value [ 556.655189] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 556.655189] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] updated_port = self._update_port( [ 556.655189] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.655189] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] _ensure_no_port_binding_failure(port) [ 556.655189] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.655189] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] raise exception.PortBindingFailed(port_id=port['id']) [ 556.655189] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] nova.exception.PortBindingFailed: Binding failed for port 1339efad-11b7-403a-9a9c-9f70795493ea, please check neutron logs for more information. [ 556.655189] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] [ 556.655713] env[62600]: INFO nova.compute.manager [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Terminating instance [ 556.655713] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.304s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.655713] env[62600]: INFO nova.compute.claims [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 556.655713] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Acquiring lock "refresh_cache-da07d0e2-98bc-451c-955e-db5eaeee3c82" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.655903] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Acquired lock "refresh_cache-da07d0e2-98bc-451c-955e-db5eaeee3c82" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.655903] env[62600]: DEBUG nova.network.neutron [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 556.747514] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.094646] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222183, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025887} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.094897] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 557.096673] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd769a74-15f4-4ab1-ba03-38fe62bd0484 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.103682] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 557.103682] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52de9a68-b73e-9e2c-df36-af59e692b29c" [ 557.103682] env[62600]: _type = "Task" [ 557.103682] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.116706] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52de9a68-b73e-9e2c-df36-af59e692b29c, 'name': SearchDatastore_Task, 'duration_secs': 0.009276} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.116706] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.116706] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 98cbcc28-54e0-40de-9feb-6da917806423/98cbcc28-54e0-40de-9feb-6da917806423.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 557.116960] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b6da95c9-22f7-428f-85fc-f5d0800156fa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.124118] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 557.124118] env[62600]: value = "task-1222184" [ 557.124118] env[62600]: _type = "Task" [ 557.124118] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.132308] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222184, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.153914] env[62600]: DEBUG nova.compute.utils [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 557.155709] env[62600]: DEBUG nova.compute.manager [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 557.156048] env[62600]: DEBUG nova.network.neutron [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 557.225625] env[62600]: DEBUG nova.network.neutron [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 557.265756] env[62600]: DEBUG nova.policy [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8bf69f87ddaf469aad86b9bdb950b85a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fc189a4692be4b50b290384dd4af1e31', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 557.321801] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquiring lock "b569fb64-27c2-4e20-a2c4-c81058958f98" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.322158] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Lock "b569fb64-27c2-4e20-a2c4-c81058958f98" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.351184] env[62600]: DEBUG nova.compute.manager [req-f5386650-e46d-42ad-bb5c-8e343eab4f9b req-a1ef5390-8a71-47ba-96b2-3d97cd5a9207 service nova] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Received event network-changed-1339efad-11b7-403a-9a9c-9f70795493ea {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 557.351428] env[62600]: DEBUG nova.compute.manager [req-f5386650-e46d-42ad-bb5c-8e343eab4f9b req-a1ef5390-8a71-47ba-96b2-3d97cd5a9207 service nova] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Refreshing instance network info cache due to event network-changed-1339efad-11b7-403a-9a9c-9f70795493ea. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 557.351893] env[62600]: DEBUG oslo_concurrency.lockutils [req-f5386650-e46d-42ad-bb5c-8e343eab4f9b req-a1ef5390-8a71-47ba-96b2-3d97cd5a9207 service nova] Acquiring lock "refresh_cache-da07d0e2-98bc-451c-955e-db5eaeee3c82" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.414773] env[62600]: DEBUG nova.network.neutron [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.556262] env[62600]: ERROR nova.compute.manager [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 92c2ae25-d1b4-47d0-a419-86065fc4d35c, please check neutron logs for more information. [ 557.556262] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 557.556262] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.556262] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 557.556262] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 557.556262] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 557.556262] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 557.556262] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 557.556262] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.556262] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 557.556262] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.556262] env[62600]: ERROR nova.compute.manager raise self.value [ 557.556262] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 557.556262] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 557.556262] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.556262] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 557.557434] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.557434] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 557.557434] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 92c2ae25-d1b4-47d0-a419-86065fc4d35c, please check neutron logs for more information. [ 557.557434] env[62600]: ERROR nova.compute.manager [ 557.557434] env[62600]: Traceback (most recent call last): [ 557.557434] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 557.557434] env[62600]: listener.cb(fileno) [ 557.557434] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.557434] env[62600]: result = function(*args, **kwargs) [ 557.557434] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 557.557434] env[62600]: return func(*args, **kwargs) [ 557.557434] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.557434] env[62600]: raise e [ 557.557434] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.557434] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 557.557434] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 557.557434] env[62600]: created_port_ids = self._update_ports_for_instance( [ 557.557434] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 557.557434] env[62600]: with excutils.save_and_reraise_exception(): [ 557.557434] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.557434] env[62600]: self.force_reraise() [ 557.557434] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.557434] env[62600]: raise self.value [ 557.557434] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 557.557434] env[62600]: updated_port = self._update_port( [ 557.557434] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.557434] env[62600]: _ensure_no_port_binding_failure(port) [ 557.557434] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.557434] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 557.559991] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 92c2ae25-d1b4-47d0-a419-86065fc4d35c, please check neutron logs for more information. [ 557.559991] env[62600]: Removing descriptor: 15 [ 557.559991] env[62600]: ERROR nova.compute.manager [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 92c2ae25-d1b4-47d0-a419-86065fc4d35c, please check neutron logs for more information. [ 557.559991] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Traceback (most recent call last): [ 557.559991] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 557.559991] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] yield resources [ 557.559991] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 557.559991] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] self.driver.spawn(context, instance, image_meta, [ 557.559991] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 557.559991] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] self._vmops.spawn(context, instance, image_meta, injected_files, [ 557.559991] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 557.559991] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] vm_ref = self.build_virtual_machine(instance, [ 557.560323] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 557.560323] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] vif_infos = vmwarevif.get_vif_info(self._session, [ 557.560323] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 557.560323] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] for vif in network_info: [ 557.560323] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 557.560323] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] return self._sync_wrapper(fn, *args, **kwargs) [ 557.560323] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 557.560323] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] self.wait() [ 557.560323] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 557.560323] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] self[:] = self._gt.wait() [ 557.560323] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 557.560323] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] return self._exit_event.wait() [ 557.560323] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 557.560648] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] result = hub.switch() [ 557.560648] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 557.560648] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] return self.greenlet.switch() [ 557.560648] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.560648] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] result = function(*args, **kwargs) [ 557.560648] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 557.560648] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] return func(*args, **kwargs) [ 557.560648] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.560648] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] raise e [ 557.560648] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.560648] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] nwinfo = self.network_api.allocate_for_instance( [ 557.560648] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 557.560648] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] created_port_ids = self._update_ports_for_instance( [ 557.560966] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 557.560966] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] with excutils.save_and_reraise_exception(): [ 557.560966] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.560966] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] self.force_reraise() [ 557.560966] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.560966] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] raise self.value [ 557.560966] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 557.560966] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] updated_port = self._update_port( [ 557.560966] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.560966] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] _ensure_no_port_binding_failure(port) [ 557.560966] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.560966] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] raise exception.PortBindingFailed(port_id=port['id']) [ 557.561259] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] nova.exception.PortBindingFailed: Binding failed for port 92c2ae25-d1b4-47d0-a419-86065fc4d35c, please check neutron logs for more information. [ 557.561259] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] [ 557.561259] env[62600]: INFO nova.compute.manager [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Terminating instance [ 557.561259] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Acquiring lock "refresh_cache-b315463b-f8aa-4b16-a5a2-069458f15358" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.561259] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Acquired lock "refresh_cache-b315463b-f8aa-4b16-a5a2-069458f15358" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.561259] env[62600]: DEBUG nova.network.neutron [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 557.643105] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222184, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.517392} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.643352] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 98cbcc28-54e0-40de-9feb-6da917806423/98cbcc28-54e0-40de-9feb-6da917806423.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 557.643478] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 557.644148] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-02962d6d-08e3-456c-85cb-e70aab887e71 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.651850] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 557.651850] env[62600]: value = "task-1222185" [ 557.651850] env[62600]: _type = "Task" [ 557.651850] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.664381] env[62600]: DEBUG nova.compute.manager [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 557.672668] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222185, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.921220] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Releasing lock "refresh_cache-da07d0e2-98bc-451c-955e-db5eaeee3c82" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.925018] env[62600]: DEBUG nova.compute.manager [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 557.925018] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 557.925903] env[62600]: DEBUG oslo_concurrency.lockutils [req-f5386650-e46d-42ad-bb5c-8e343eab4f9b req-a1ef5390-8a71-47ba-96b2-3d97cd5a9207 service nova] Acquired lock "refresh_cache-da07d0e2-98bc-451c-955e-db5eaeee3c82" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.926131] env[62600]: DEBUG nova.network.neutron [req-f5386650-e46d-42ad-bb5c-8e343eab4f9b req-a1ef5390-8a71-47ba-96b2-3d97cd5a9207 service nova] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Refreshing network info cache for port 1339efad-11b7-403a-9a9c-9f70795493ea {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 557.927885] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fa83ac10-bb65-4320-b66a-d00038f42fd6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.944258] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4629bd5b-f842-4c09-b191-c99f29d3782c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.979236] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance da07d0e2-98bc-451c-955e-db5eaeee3c82 could not be found. [ 557.979572] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 557.979885] env[62600]: INFO nova.compute.manager [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Took 0.06 seconds to destroy the instance on the hypervisor. [ 557.980169] env[62600]: DEBUG oslo.service.loopingcall [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 557.980390] env[62600]: DEBUG nova.compute.manager [-] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 557.980485] env[62600]: DEBUG nova.network.neutron [-] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 558.021692] env[62600]: DEBUG nova.network.neutron [-] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 558.029475] env[62600]: DEBUG nova.network.neutron [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Successfully created port: a9a15762-111c-4f9c-bd35-b106aec57e3b {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 558.053559] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cead42d-b6c2-4893-b116-83dbaf6039d3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.063196] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f21f736-56b9-43d2-8ece-dec6cd4d2755 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.096208] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9abba69b-014c-4297-b8b4-32b6b89c041c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.105943] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a22c68f-464e-4ec7-a4f2-34858f068077 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.113262] env[62600]: DEBUG nova.network.neutron [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 558.125354] env[62600]: DEBUG nova.compute.provider_tree [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 558.165655] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222185, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063851} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.165980] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 558.166880] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a97d2fc-5564-4ed2-9703-39aa6648daad {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.198664] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Reconfiguring VM instance instance-00000001 to attach disk [datastore2] 98cbcc28-54e0-40de-9feb-6da917806423/98cbcc28-54e0-40de-9feb-6da917806423.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 558.198664] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dfda1dbc-a96b-4b86-95ce-f702d8b066c4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.220807] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 558.220807] env[62600]: value = "task-1222186" [ 558.220807] env[62600]: _type = "Task" [ 558.220807] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.233723] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222186, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.234505] env[62600]: DEBUG nova.network.neutron [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.465394] env[62600]: DEBUG nova.network.neutron [req-f5386650-e46d-42ad-bb5c-8e343eab4f9b req-a1ef5390-8a71-47ba-96b2-3d97cd5a9207 service nova] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 558.524977] env[62600]: DEBUG nova.network.neutron [-] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.550541] env[62600]: DEBUG nova.network.neutron [req-f5386650-e46d-42ad-bb5c-8e343eab4f9b req-a1ef5390-8a71-47ba-96b2-3d97cd5a9207 service nova] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.629664] env[62600]: DEBUG nova.scheduler.client.report [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 558.687125] env[62600]: DEBUG nova.compute.manager [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 558.713919] env[62600]: DEBUG nova.virt.hardware [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 558.714160] env[62600]: DEBUG nova.virt.hardware [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 558.714324] env[62600]: DEBUG nova.virt.hardware [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 558.714510] env[62600]: DEBUG nova.virt.hardware [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 558.714653] env[62600]: DEBUG nova.virt.hardware [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 558.714795] env[62600]: DEBUG nova.virt.hardware [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 558.714989] env[62600]: DEBUG nova.virt.hardware [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 558.715464] env[62600]: DEBUG nova.virt.hardware [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 558.716074] env[62600]: DEBUG nova.virt.hardware [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 558.716074] env[62600]: DEBUG nova.virt.hardware [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 558.717020] env[62600]: DEBUG nova.virt.hardware [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 558.718536] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4abd5290-4761-43b5-8704-47ee20a27ebc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.733434] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-162060a5-c89e-4aa6-beb0-46d07ef6f338 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.737613] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Releasing lock "refresh_cache-b315463b-f8aa-4b16-a5a2-069458f15358" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.738038] env[62600]: DEBUG nova.compute.manager [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 558.738228] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 558.739340] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3df5c881-224b-489d-a91f-e0a50adbee5e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.756960] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222186, 'name': ReconfigVM_Task, 'duration_secs': 0.278097} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.759343] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Reconfigured VM instance instance-00000001 to attach disk [datastore2] 98cbcc28-54e0-40de-9feb-6da917806423/98cbcc28-54e0-40de-9feb-6da917806423.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 558.759794] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f0b4bf6c-af8f-4fc5-9f5d-f7a3f3b41840 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.772604] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 558.772604] env[62600]: value = "task-1222187" [ 558.772604] env[62600]: _type = "Task" [ 558.772604] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.775582] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-498b82cc-0c73-489b-a204-cade41712260 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.799953] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222187, 'name': Rename_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.803855] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b315463b-f8aa-4b16-a5a2-069458f15358 could not be found. [ 558.804041] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 558.804924] env[62600]: INFO nova.compute.manager [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Took 0.07 seconds to destroy the instance on the hypervisor. [ 558.804924] env[62600]: DEBUG oslo.service.loopingcall [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 558.804924] env[62600]: DEBUG nova.compute.manager [-] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 558.804924] env[62600]: DEBUG nova.network.neutron [-] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 558.827902] env[62600]: DEBUG nova.network.neutron [-] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 559.004943] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquiring lock "dc77594b-7767-40bf-adbb-c72fa5c80503" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.004943] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Lock "dc77594b-7767-40bf-adbb-c72fa5c80503" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.032525] env[62600]: INFO nova.compute.manager [-] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Took 1.05 seconds to deallocate network for instance. [ 559.035644] env[62600]: DEBUG nova.compute.claims [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 559.035862] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.055668] env[62600]: DEBUG oslo_concurrency.lockutils [req-f5386650-e46d-42ad-bb5c-8e343eab4f9b req-a1ef5390-8a71-47ba-96b2-3d97cd5a9207 service nova] Releasing lock "refresh_cache-da07d0e2-98bc-451c-955e-db5eaeee3c82" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.099832] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Acquiring lock "8e64b614-31e3-4830-b620-ff8409da8c37" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.100081] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Lock "8e64b614-31e3-4830-b620-ff8409da8c37" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.135179] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.135711] env[62600]: DEBUG nova.compute.manager [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 559.139786] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.590s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.142894] env[62600]: INFO nova.compute.claims [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 559.298554] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222187, 'name': Rename_Task, 'duration_secs': 0.1423} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.298824] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 559.299076] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef003c5a-51cd-43ce-af19-602e2d85fbe9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.305576] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 559.305576] env[62600]: value = "task-1222188" [ 559.305576] env[62600]: _type = "Task" [ 559.305576] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.318912] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222188, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.330468] env[62600]: DEBUG nova.network.neutron [-] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.426217] env[62600]: ERROR nova.compute.manager [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a9a15762-111c-4f9c-bd35-b106aec57e3b, please check neutron logs for more information. [ 559.426217] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 559.426217] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.426217] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 559.426217] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 559.426217] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 559.426217] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 559.426217] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 559.426217] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.426217] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 559.426217] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.426217] env[62600]: ERROR nova.compute.manager raise self.value [ 559.426217] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 559.426217] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 559.426217] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.426217] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 559.426855] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.426855] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 559.426855] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a9a15762-111c-4f9c-bd35-b106aec57e3b, please check neutron logs for more information. [ 559.426855] env[62600]: ERROR nova.compute.manager [ 559.426855] env[62600]: Traceback (most recent call last): [ 559.426855] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 559.426855] env[62600]: listener.cb(fileno) [ 559.426855] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.426855] env[62600]: result = function(*args, **kwargs) [ 559.426855] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.426855] env[62600]: return func(*args, **kwargs) [ 559.426855] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 559.426855] env[62600]: raise e [ 559.426855] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.426855] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 559.426855] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 559.426855] env[62600]: created_port_ids = self._update_ports_for_instance( [ 559.426855] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 559.426855] env[62600]: with excutils.save_and_reraise_exception(): [ 559.426855] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.426855] env[62600]: self.force_reraise() [ 559.426855] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.426855] env[62600]: raise self.value [ 559.426855] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 559.426855] env[62600]: updated_port = self._update_port( [ 559.426855] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.426855] env[62600]: _ensure_no_port_binding_failure(port) [ 559.426855] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.426855] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 559.427898] env[62600]: nova.exception.PortBindingFailed: Binding failed for port a9a15762-111c-4f9c-bd35-b106aec57e3b, please check neutron logs for more information. [ 559.427898] env[62600]: Removing descriptor: 16 [ 559.427898] env[62600]: ERROR nova.compute.manager [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a9a15762-111c-4f9c-bd35-b106aec57e3b, please check neutron logs for more information. [ 559.427898] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] Traceback (most recent call last): [ 559.427898] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 559.427898] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] yield resources [ 559.427898] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 559.427898] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] self.driver.spawn(context, instance, image_meta, [ 559.427898] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 559.427898] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 559.427898] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 559.427898] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] vm_ref = self.build_virtual_machine(instance, [ 559.428214] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 559.428214] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] vif_infos = vmwarevif.get_vif_info(self._session, [ 559.428214] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 559.428214] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] for vif in network_info: [ 559.428214] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 559.428214] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] return self._sync_wrapper(fn, *args, **kwargs) [ 559.428214] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 559.428214] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] self.wait() [ 559.428214] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 559.428214] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] self[:] = self._gt.wait() [ 559.428214] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 559.428214] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] return self._exit_event.wait() [ 559.428214] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 559.428571] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] result = hub.switch() [ 559.428571] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 559.428571] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] return self.greenlet.switch() [ 559.428571] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.428571] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] result = function(*args, **kwargs) [ 559.428571] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.428571] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] return func(*args, **kwargs) [ 559.428571] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 559.428571] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] raise e [ 559.428571] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.428571] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] nwinfo = self.network_api.allocate_for_instance( [ 559.428571] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 559.428571] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] created_port_ids = self._update_ports_for_instance( [ 559.428912] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 559.428912] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] with excutils.save_and_reraise_exception(): [ 559.428912] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.428912] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] self.force_reraise() [ 559.428912] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.428912] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] raise self.value [ 559.428912] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 559.428912] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] updated_port = self._update_port( [ 559.428912] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.428912] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] _ensure_no_port_binding_failure(port) [ 559.428912] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.428912] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] raise exception.PortBindingFailed(port_id=port['id']) [ 559.429229] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] nova.exception.PortBindingFailed: Binding failed for port a9a15762-111c-4f9c-bd35-b106aec57e3b, please check neutron logs for more information. [ 559.429229] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] [ 559.429229] env[62600]: INFO nova.compute.manager [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Terminating instance [ 559.431168] env[62600]: DEBUG oslo_concurrency.lockutils [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Acquiring lock "refresh_cache-371a3210-1741-43ce-b635-b84314aee77e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.431168] env[62600]: DEBUG oslo_concurrency.lockutils [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Acquired lock "refresh_cache-371a3210-1741-43ce-b635-b84314aee77e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.431278] env[62600]: DEBUG nova.network.neutron [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 559.648807] env[62600]: DEBUG nova.compute.utils [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 559.651408] env[62600]: DEBUG nova.compute.manager [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 559.651614] env[62600]: DEBUG nova.network.neutron [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 559.816130] env[62600]: DEBUG oslo_vmware.api [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222188, 'name': PowerOnVM_Task, 'duration_secs': 0.423473} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.816963] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 559.817398] env[62600]: INFO nova.compute.manager [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Took 8.44 seconds to spawn the instance on the hypervisor. [ 559.820083] env[62600]: DEBUG nova.compute.manager [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 559.820083] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb61a42-11e3-4699-bdf9-c6714d9623eb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.840904] env[62600]: INFO nova.compute.manager [-] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Took 1.04 seconds to deallocate network for instance. [ 559.849174] env[62600]: DEBUG nova.compute.claims [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 559.849174] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.889948] env[62600]: DEBUG nova.policy [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '749181214438458790664c5ac9ab5575', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6ce86d7fc7bf411bbddc1a8bc1b35963', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 559.958818] env[62600]: DEBUG nova.network.neutron [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 560.041665] env[62600]: DEBUG nova.network.neutron [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.152612] env[62600]: DEBUG nova.compute.manager [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 560.297589] env[62600]: DEBUG nova.network.neutron [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Successfully created port: be017d6a-89b0-49f0-8149-7e3eb6b91ef9 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 560.348341] env[62600]: INFO nova.compute.manager [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Took 13.29 seconds to build instance. [ 560.515508] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6700165-6e05-48e4-b31d-d213231444dd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.526659] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf168f5-3cb5-498d-a959-9b0339280ceb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.559168] env[62600]: DEBUG oslo_concurrency.lockutils [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Releasing lock "refresh_cache-371a3210-1741-43ce-b635-b84314aee77e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.559588] env[62600]: DEBUG nova.compute.manager [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 560.559792] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 560.560382] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6869f250-5c15-4076-bc56-9bd1ed5647b5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.563452] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c39356-995f-483f-a3a3-559b2c2abd7b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.572176] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a21d600-b700-4ac2-b0cc-3cb2c6ea9977 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.578741] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b9b9fad-77c3-415e-96c5-07a7ce67a1c5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.599922] env[62600]: DEBUG nova.compute.provider_tree [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 560.605234] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 371a3210-1741-43ce-b635-b84314aee77e could not be found. [ 560.605507] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 560.605584] env[62600]: INFO nova.compute.manager [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Took 0.05 seconds to destroy the instance on the hypervisor. [ 560.605822] env[62600]: DEBUG oslo.service.loopingcall [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 560.606271] env[62600]: DEBUG nova.compute.manager [-] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 560.606375] env[62600]: DEBUG nova.network.neutron [-] [instance: 371a3210-1741-43ce-b635-b84314aee77e] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 560.626585] env[62600]: DEBUG nova.network.neutron [-] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 560.852173] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e81a75f2-5d71-475d-b513-009cf0f88bc8 tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Lock "98cbcc28-54e0-40de-9feb-6da917806423" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.813s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.036138] env[62600]: DEBUG nova.compute.manager [req-3cf0f34e-7bb1-4a13-ac0c-ae60b43388d7 req-a2b41352-b646-4d74-a5c1-4c91cd78178a service nova] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Received event network-vif-deleted-1339efad-11b7-403a-9a9c-9f70795493ea {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 561.036358] env[62600]: DEBUG nova.compute.manager [req-3cf0f34e-7bb1-4a13-ac0c-ae60b43388d7 req-a2b41352-b646-4d74-a5c1-4c91cd78178a service nova] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Received event network-changed-92c2ae25-d1b4-47d0-a419-86065fc4d35c {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 561.036521] env[62600]: DEBUG nova.compute.manager [req-3cf0f34e-7bb1-4a13-ac0c-ae60b43388d7 req-a2b41352-b646-4d74-a5c1-4c91cd78178a service nova] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Refreshing instance network info cache due to event network-changed-92c2ae25-d1b4-47d0-a419-86065fc4d35c. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 561.036729] env[62600]: DEBUG oslo_concurrency.lockutils [req-3cf0f34e-7bb1-4a13-ac0c-ae60b43388d7 req-a2b41352-b646-4d74-a5c1-4c91cd78178a service nova] Acquiring lock "refresh_cache-b315463b-f8aa-4b16-a5a2-069458f15358" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.036866] env[62600]: DEBUG oslo_concurrency.lockutils [req-3cf0f34e-7bb1-4a13-ac0c-ae60b43388d7 req-a2b41352-b646-4d74-a5c1-4c91cd78178a service nova] Acquired lock "refresh_cache-b315463b-f8aa-4b16-a5a2-069458f15358" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.037940] env[62600]: DEBUG nova.network.neutron [req-3cf0f34e-7bb1-4a13-ac0c-ae60b43388d7 req-a2b41352-b646-4d74-a5c1-4c91cd78178a service nova] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Refreshing network info cache for port 92c2ae25-d1b4-47d0-a419-86065fc4d35c {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 561.113254] env[62600]: DEBUG nova.scheduler.client.report [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 561.129128] env[62600]: DEBUG nova.network.neutron [-] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.170864] env[62600]: DEBUG nova.compute.manager [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 561.207893] env[62600]: DEBUG nova.virt.hardware [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 561.207893] env[62600]: DEBUG nova.virt.hardware [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 561.207893] env[62600]: DEBUG nova.virt.hardware [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 561.209067] env[62600]: DEBUG nova.virt.hardware [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 561.209067] env[62600]: DEBUG nova.virt.hardware [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 561.209067] env[62600]: DEBUG nova.virt.hardware [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 561.209067] env[62600]: DEBUG nova.virt.hardware [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 561.209067] env[62600]: DEBUG nova.virt.hardware [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 561.209459] env[62600]: DEBUG nova.virt.hardware [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 561.209459] env[62600]: DEBUG nova.virt.hardware [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 561.209459] env[62600]: DEBUG nova.virt.hardware [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 561.209459] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f418ff-aeec-4e98-914a-e1e2aefebe16 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.223382] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ffd80a1-9f41-4e83-99e6-5e1503a4ea37 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.357706] env[62600]: DEBUG nova.compute.manager [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 561.366627] env[62600]: ERROR nova.compute.manager [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port be017d6a-89b0-49f0-8149-7e3eb6b91ef9, please check neutron logs for more information. [ 561.366627] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 561.366627] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 561.366627] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 561.366627] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 561.366627] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 561.366627] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 561.366627] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 561.366627] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.366627] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 561.366627] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.366627] env[62600]: ERROR nova.compute.manager raise self.value [ 561.366627] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 561.366627] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 561.366627] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.366627] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 561.367496] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.367496] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 561.367496] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port be017d6a-89b0-49f0-8149-7e3eb6b91ef9, please check neutron logs for more information. [ 561.367496] env[62600]: ERROR nova.compute.manager [ 561.367496] env[62600]: Traceback (most recent call last): [ 561.367496] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 561.367496] env[62600]: listener.cb(fileno) [ 561.367496] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.367496] env[62600]: result = function(*args, **kwargs) [ 561.367496] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 561.367496] env[62600]: return func(*args, **kwargs) [ 561.367496] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 561.367496] env[62600]: raise e [ 561.367496] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 561.367496] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 561.367496] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 561.367496] env[62600]: created_port_ids = self._update_ports_for_instance( [ 561.367496] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 561.367496] env[62600]: with excutils.save_and_reraise_exception(): [ 561.367496] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.367496] env[62600]: self.force_reraise() [ 561.367496] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.367496] env[62600]: raise self.value [ 561.367496] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 561.367496] env[62600]: updated_port = self._update_port( [ 561.367496] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.367496] env[62600]: _ensure_no_port_binding_failure(port) [ 561.367496] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.367496] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 561.369032] env[62600]: nova.exception.PortBindingFailed: Binding failed for port be017d6a-89b0-49f0-8149-7e3eb6b91ef9, please check neutron logs for more information. [ 561.369032] env[62600]: Removing descriptor: 16 [ 561.369032] env[62600]: ERROR nova.compute.manager [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port be017d6a-89b0-49f0-8149-7e3eb6b91ef9, please check neutron logs for more information. [ 561.369032] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Traceback (most recent call last): [ 561.369032] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 561.369032] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] yield resources [ 561.369032] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 561.369032] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] self.driver.spawn(context, instance, image_meta, [ 561.369032] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 561.369032] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 561.369032] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 561.369032] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] vm_ref = self.build_virtual_machine(instance, [ 561.370106] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 561.370106] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] vif_infos = vmwarevif.get_vif_info(self._session, [ 561.370106] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 561.370106] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] for vif in network_info: [ 561.370106] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 561.370106] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] return self._sync_wrapper(fn, *args, **kwargs) [ 561.370106] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 561.370106] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] self.wait() [ 561.370106] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 561.370106] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] self[:] = self._gt.wait() [ 561.370106] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 561.370106] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] return self._exit_event.wait() [ 561.370106] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 561.370753] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] result = hub.switch() [ 561.370753] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 561.370753] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] return self.greenlet.switch() [ 561.370753] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.370753] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] result = function(*args, **kwargs) [ 561.370753] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 561.370753] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] return func(*args, **kwargs) [ 561.370753] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 561.370753] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] raise e [ 561.370753] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 561.370753] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] nwinfo = self.network_api.allocate_for_instance( [ 561.370753] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 561.370753] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] created_port_ids = self._update_ports_for_instance( [ 561.372121] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 561.372121] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] with excutils.save_and_reraise_exception(): [ 561.372121] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.372121] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] self.force_reraise() [ 561.372121] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.372121] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] raise self.value [ 561.372121] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 561.372121] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] updated_port = self._update_port( [ 561.372121] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.372121] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] _ensure_no_port_binding_failure(port) [ 561.372121] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.372121] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] raise exception.PortBindingFailed(port_id=port['id']) [ 561.372525] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] nova.exception.PortBindingFailed: Binding failed for port be017d6a-89b0-49f0-8149-7e3eb6b91ef9, please check neutron logs for more information. [ 561.372525] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] [ 561.372525] env[62600]: INFO nova.compute.manager [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Terminating instance [ 561.372525] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Acquiring lock "refresh_cache-12d03944-78d1-44e4-aaf9-5678f7f3e1ea" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.372525] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Acquired lock "refresh_cache-12d03944-78d1-44e4-aaf9-5678f7f3e1ea" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.372525] env[62600]: DEBUG nova.network.neutron [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 561.562440] env[62600]: DEBUG nova.network.neutron [req-3cf0f34e-7bb1-4a13-ac0c-ae60b43388d7 req-a2b41352-b646-4d74-a5c1-4c91cd78178a service nova] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 561.617164] env[62600]: DEBUG nova.network.neutron [req-3cf0f34e-7bb1-4a13-ac0c-ae60b43388d7 req-a2b41352-b646-4d74-a5c1-4c91cd78178a service nova] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.618753] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.481s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.619222] env[62600]: DEBUG nova.compute.manager [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 561.625517] env[62600]: DEBUG oslo_concurrency.lockutils [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.206s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.625517] env[62600]: INFO nova.compute.claims [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 561.629274] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquiring lock "6467456c-db39-4fd6-b67e-a5be2b803bd3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.629483] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Lock "6467456c-db39-4fd6-b67e-a5be2b803bd3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.632120] env[62600]: INFO nova.compute.manager [-] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Took 1.03 seconds to deallocate network for instance. [ 561.633945] env[62600]: DEBUG nova.compute.claims [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 561.634121] env[62600]: DEBUG oslo_concurrency.lockutils [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.879659] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.897633] env[62600]: DEBUG nova.network.neutron [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 561.948885] env[62600]: DEBUG nova.network.neutron [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.122642] env[62600]: DEBUG oslo_concurrency.lockutils [req-3cf0f34e-7bb1-4a13-ac0c-ae60b43388d7 req-a2b41352-b646-4d74-a5c1-4c91cd78178a service nova] Releasing lock "refresh_cache-b315463b-f8aa-4b16-a5a2-069458f15358" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.123035] env[62600]: DEBUG nova.compute.manager [req-3cf0f34e-7bb1-4a13-ac0c-ae60b43388d7 req-a2b41352-b646-4d74-a5c1-4c91cd78178a service nova] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Received event network-vif-deleted-92c2ae25-d1b4-47d0-a419-86065fc4d35c {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 562.123400] env[62600]: DEBUG nova.compute.manager [req-3cf0f34e-7bb1-4a13-ac0c-ae60b43388d7 req-a2b41352-b646-4d74-a5c1-4c91cd78178a service nova] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Received event network-changed-a9a15762-111c-4f9c-bd35-b106aec57e3b {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 562.123835] env[62600]: DEBUG nova.compute.manager [req-3cf0f34e-7bb1-4a13-ac0c-ae60b43388d7 req-a2b41352-b646-4d74-a5c1-4c91cd78178a service nova] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Refreshing instance network info cache due to event network-changed-a9a15762-111c-4f9c-bd35-b106aec57e3b. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 562.124207] env[62600]: DEBUG oslo_concurrency.lockutils [req-3cf0f34e-7bb1-4a13-ac0c-ae60b43388d7 req-a2b41352-b646-4d74-a5c1-4c91cd78178a service nova] Acquiring lock "refresh_cache-371a3210-1741-43ce-b635-b84314aee77e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.124558] env[62600]: DEBUG oslo_concurrency.lockutils [req-3cf0f34e-7bb1-4a13-ac0c-ae60b43388d7 req-a2b41352-b646-4d74-a5c1-4c91cd78178a service nova] Acquired lock "refresh_cache-371a3210-1741-43ce-b635-b84314aee77e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.124748] env[62600]: DEBUG nova.network.neutron [req-3cf0f34e-7bb1-4a13-ac0c-ae60b43388d7 req-a2b41352-b646-4d74-a5c1-4c91cd78178a service nova] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Refreshing network info cache for port a9a15762-111c-4f9c-bd35-b106aec57e3b {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 562.127019] env[62600]: DEBUG nova.compute.utils [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 562.130245] env[62600]: DEBUG nova.compute.manager [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 562.130245] env[62600]: DEBUG nova.network.neutron [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 562.185456] env[62600]: DEBUG nova.policy [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7c3484585a43467cb89bb169071ae657', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '88c4dc2bccfd4d72bfaeb6916142b19f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 562.453096] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Releasing lock "refresh_cache-12d03944-78d1-44e4-aaf9-5678f7f3e1ea" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.453665] env[62600]: DEBUG nova.compute.manager [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 562.454144] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 562.454982] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Acquiring lock "2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.456194] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Lock "2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.456194] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-87edb8b8-478b-4edc-adef-93fdd7f5ece4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.466085] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cab13ed-a198-44cc-9fed-0d395877611f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.481436] env[62600]: DEBUG nova.network.neutron [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Successfully created port: a10a7f80-347e-49c9-aae5-76ca002d1a38 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 562.497731] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 12d03944-78d1-44e4-aaf9-5678f7f3e1ea could not be found. [ 562.497852] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 562.498082] env[62600]: INFO nova.compute.manager [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Took 0.04 seconds to destroy the instance on the hypervisor. [ 562.498289] env[62600]: DEBUG oslo.service.loopingcall [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 562.498544] env[62600]: DEBUG nova.compute.manager [-] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 562.498632] env[62600]: DEBUG nova.network.neutron [-] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 562.527259] env[62600]: DEBUG nova.network.neutron [-] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 562.634388] env[62600]: DEBUG nova.compute.manager [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 562.652443] env[62600]: DEBUG nova.network.neutron [req-3cf0f34e-7bb1-4a13-ac0c-ae60b43388d7 req-a2b41352-b646-4d74-a5c1-4c91cd78178a service nova] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 562.727286] env[62600]: DEBUG nova.network.neutron [req-3cf0f34e-7bb1-4a13-ac0c-ae60b43388d7 req-a2b41352-b646-4d74-a5c1-4c91cd78178a service nova] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.984697] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5edec55-2e50-4466-8dfa-2aca546fe718 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.992820] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9854d2a-e9b4-4bb1-9a1c-40274c96b295 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.025206] env[62600]: INFO nova.compute.manager [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Rebuilding instance [ 563.028389] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc771cfb-276a-43c7-9061-b46238722eac {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.031607] env[62600]: DEBUG nova.network.neutron [-] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.038536] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b019d03b-ec5f-49d9-bf42-1698243aed9a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.054073] env[62600]: DEBUG nova.compute.provider_tree [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 563.088898] env[62600]: DEBUG nova.compute.manager [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 563.089937] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f727f471-c414-45b5-bd70-50b9cc4a8d90 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.229538] env[62600]: DEBUG oslo_concurrency.lockutils [req-3cf0f34e-7bb1-4a13-ac0c-ae60b43388d7 req-a2b41352-b646-4d74-a5c1-4c91cd78178a service nova] Releasing lock "refresh_cache-371a3210-1741-43ce-b635-b84314aee77e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.535100] env[62600]: INFO nova.compute.manager [-] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Took 1.04 seconds to deallocate network for instance. [ 563.537950] env[62600]: DEBUG nova.compute.claims [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 563.538204] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.558907] env[62600]: DEBUG nova.scheduler.client.report [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 563.602902] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 563.603551] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7cabb31a-2c61-40d8-9c7b-54f85f4440a4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.612449] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 563.612449] env[62600]: value = "task-1222189" [ 563.612449] env[62600]: _type = "Task" [ 563.612449] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.626982] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222189, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.648293] env[62600]: DEBUG nova.compute.manager [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 563.683498] env[62600]: DEBUG nova.virt.hardware [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 563.683498] env[62600]: DEBUG nova.virt.hardware [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 563.683498] env[62600]: DEBUG nova.virt.hardware [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 563.684206] env[62600]: DEBUG nova.virt.hardware [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 563.684416] env[62600]: DEBUG nova.virt.hardware [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 563.684575] env[62600]: DEBUG nova.virt.hardware [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 563.684787] env[62600]: DEBUG nova.virt.hardware [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 563.685219] env[62600]: DEBUG nova.virt.hardware [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 563.685678] env[62600]: DEBUG nova.virt.hardware [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 563.686037] env[62600]: DEBUG nova.virt.hardware [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 563.686512] env[62600]: DEBUG nova.virt.hardware [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 563.687713] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c697a03d-8738-40fc-8ba9-dd8d559e153d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.697223] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cecbb9c7-1f18-4190-ae34-9b70f85c532f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.066978] env[62600]: DEBUG oslo_concurrency.lockutils [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.067586] env[62600]: DEBUG nova.compute.manager [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 564.072682] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.431s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.072982] env[62600]: INFO nova.compute.claims [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 564.130174] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222189, 'name': PowerOffVM_Task, 'duration_secs': 0.116631} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 564.130506] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 564.130814] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 564.131578] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b0c0040-e8e8-4846-84af-361fc006a5f7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.141706] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 564.141971] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2e904f56-9663-46e5-8954-5e3eac5400d8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.167620] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 564.167620] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 564.167814] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Deleting the datastore file [datastore2] 98cbcc28-54e0-40de-9feb-6da917806423 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 564.168618] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-846ce36f-899d-4d21-8877-e2d91ad7045d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.175127] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 564.175127] env[62600]: value = "task-1222191" [ 564.175127] env[62600]: _type = "Task" [ 564.175127] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.187248] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222191, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.239470] env[62600]: ERROR nova.compute.manager [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a10a7f80-347e-49c9-aae5-76ca002d1a38, please check neutron logs for more information. [ 564.239470] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 564.239470] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 564.239470] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 564.239470] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 564.239470] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 564.239470] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 564.239470] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 564.239470] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.239470] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 564.239470] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.239470] env[62600]: ERROR nova.compute.manager raise self.value [ 564.239470] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 564.239470] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 564.239470] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.239470] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 564.239927] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.239927] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 564.239927] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a10a7f80-347e-49c9-aae5-76ca002d1a38, please check neutron logs for more information. [ 564.239927] env[62600]: ERROR nova.compute.manager [ 564.239927] env[62600]: Traceback (most recent call last): [ 564.239927] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 564.239927] env[62600]: listener.cb(fileno) [ 564.239927] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 564.239927] env[62600]: result = function(*args, **kwargs) [ 564.239927] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 564.239927] env[62600]: return func(*args, **kwargs) [ 564.239927] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 564.239927] env[62600]: raise e [ 564.239927] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 564.239927] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 564.239927] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 564.239927] env[62600]: created_port_ids = self._update_ports_for_instance( [ 564.239927] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 564.239927] env[62600]: with excutils.save_and_reraise_exception(): [ 564.239927] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.239927] env[62600]: self.force_reraise() [ 564.239927] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.239927] env[62600]: raise self.value [ 564.239927] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 564.239927] env[62600]: updated_port = self._update_port( [ 564.239927] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.239927] env[62600]: _ensure_no_port_binding_failure(port) [ 564.239927] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.239927] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 564.240637] env[62600]: nova.exception.PortBindingFailed: Binding failed for port a10a7f80-347e-49c9-aae5-76ca002d1a38, please check neutron logs for more information. [ 564.240637] env[62600]: Removing descriptor: 16 [ 564.240637] env[62600]: ERROR nova.compute.manager [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a10a7f80-347e-49c9-aae5-76ca002d1a38, please check neutron logs for more information. [ 564.240637] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Traceback (most recent call last): [ 564.240637] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 564.240637] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] yield resources [ 564.240637] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 564.240637] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] self.driver.spawn(context, instance, image_meta, [ 564.240637] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 564.240637] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] self._vmops.spawn(context, instance, image_meta, injected_files, [ 564.240637] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 564.240637] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] vm_ref = self.build_virtual_machine(instance, [ 564.240965] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 564.240965] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] vif_infos = vmwarevif.get_vif_info(self._session, [ 564.240965] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 564.240965] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] for vif in network_info: [ 564.240965] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 564.240965] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] return self._sync_wrapper(fn, *args, **kwargs) [ 564.240965] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 564.240965] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] self.wait() [ 564.240965] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 564.240965] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] self[:] = self._gt.wait() [ 564.240965] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 564.240965] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] return self._exit_event.wait() [ 564.240965] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 564.241363] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] result = hub.switch() [ 564.241363] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 564.241363] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] return self.greenlet.switch() [ 564.241363] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 564.241363] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] result = function(*args, **kwargs) [ 564.241363] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 564.241363] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] return func(*args, **kwargs) [ 564.241363] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 564.241363] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] raise e [ 564.241363] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 564.241363] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] nwinfo = self.network_api.allocate_for_instance( [ 564.241363] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 564.241363] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] created_port_ids = self._update_ports_for_instance( [ 564.241702] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 564.241702] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] with excutils.save_and_reraise_exception(): [ 564.241702] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.241702] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] self.force_reraise() [ 564.241702] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.241702] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] raise self.value [ 564.241702] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 564.241702] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] updated_port = self._update_port( [ 564.241702] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.241702] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] _ensure_no_port_binding_failure(port) [ 564.241702] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.241702] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] raise exception.PortBindingFailed(port_id=port['id']) [ 564.242141] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] nova.exception.PortBindingFailed: Binding failed for port a10a7f80-347e-49c9-aae5-76ca002d1a38, please check neutron logs for more information. [ 564.242141] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] [ 564.242141] env[62600]: INFO nova.compute.manager [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Terminating instance [ 564.242477] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Acquiring lock "refresh_cache-5ebc4807-907b-4254-8496-92f1bbb7b761" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 564.242636] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Acquired lock "refresh_cache-5ebc4807-907b-4254-8496-92f1bbb7b761" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 564.242799] env[62600]: DEBUG nova.network.neutron [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 564.378078] env[62600]: DEBUG nova.compute.manager [req-8beb0e1a-3d3a-4b89-96dc-a13e55e275fd req-63757ada-885b-429d-a549-d589fb11a01a service nova] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Received event network-vif-deleted-a9a15762-111c-4f9c-bd35-b106aec57e3b {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 564.378443] env[62600]: DEBUG nova.compute.manager [req-8beb0e1a-3d3a-4b89-96dc-a13e55e275fd req-63757ada-885b-429d-a549-d589fb11a01a service nova] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Received event network-changed-be017d6a-89b0-49f0-8149-7e3eb6b91ef9 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 564.378615] env[62600]: DEBUG nova.compute.manager [req-8beb0e1a-3d3a-4b89-96dc-a13e55e275fd req-63757ada-885b-429d-a549-d589fb11a01a service nova] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Refreshing instance network info cache due to event network-changed-be017d6a-89b0-49f0-8149-7e3eb6b91ef9. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 564.378929] env[62600]: DEBUG oslo_concurrency.lockutils [req-8beb0e1a-3d3a-4b89-96dc-a13e55e275fd req-63757ada-885b-429d-a549-d589fb11a01a service nova] Acquiring lock "refresh_cache-12d03944-78d1-44e4-aaf9-5678f7f3e1ea" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 564.379975] env[62600]: DEBUG oslo_concurrency.lockutils [req-8beb0e1a-3d3a-4b89-96dc-a13e55e275fd req-63757ada-885b-429d-a549-d589fb11a01a service nova] Acquired lock "refresh_cache-12d03944-78d1-44e4-aaf9-5678f7f3e1ea" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 564.379975] env[62600]: DEBUG nova.network.neutron [req-8beb0e1a-3d3a-4b89-96dc-a13e55e275fd req-63757ada-885b-429d-a549-d589fb11a01a service nova] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Refreshing network info cache for port be017d6a-89b0-49f0-8149-7e3eb6b91ef9 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 564.572750] env[62600]: DEBUG nova.compute.utils [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 564.574612] env[62600]: DEBUG nova.compute.manager [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 564.574612] env[62600]: DEBUG nova.network.neutron [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 564.656752] env[62600]: DEBUG nova.policy [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34319d68370c48ef93fa24a9d9132407', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bab74e3ae78248909dbcd483abbb8da2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 564.685821] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222191, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.118904} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 564.685968] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 564.686258] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 564.686591] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 564.884733] env[62600]: DEBUG nova.network.neutron [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 564.923513] env[62600]: DEBUG nova.network.neutron [req-8beb0e1a-3d3a-4b89-96dc-a13e55e275fd req-63757ada-885b-429d-a549-d589fb11a01a service nova] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 565.066674] env[62600]: DEBUG nova.network.neutron [req-8beb0e1a-3d3a-4b89-96dc-a13e55e275fd req-63757ada-885b-429d-a549-d589fb11a01a service nova] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.079248] env[62600]: DEBUG nova.compute.manager [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 565.149554] env[62600]: DEBUG nova.network.neutron [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.339149] env[62600]: DEBUG nova.network.neutron [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Successfully created port: 987155ef-8896-4776-83a8-4ac68b0e1f18 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 565.440786] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7059de09-cb95-46de-9094-ab1b00ba03a3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.451623] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f66eba25-3cfe-48bd-9774-8ad9d2bc2646 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.498240] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b93cdec6-23ab-4c97-8a66-f54b9b1646f1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.506800] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5776ce2-f186-4842-9fb8-3698c91647e9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.522558] env[62600]: DEBUG nova.compute.provider_tree [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.573382] env[62600]: DEBUG oslo_concurrency.lockutils [req-8beb0e1a-3d3a-4b89-96dc-a13e55e275fd req-63757ada-885b-429d-a549-d589fb11a01a service nova] Releasing lock "refresh_cache-12d03944-78d1-44e4-aaf9-5678f7f3e1ea" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.573655] env[62600]: DEBUG nova.compute.manager [req-8beb0e1a-3d3a-4b89-96dc-a13e55e275fd req-63757ada-885b-429d-a549-d589fb11a01a service nova] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Received event network-vif-deleted-be017d6a-89b0-49f0-8149-7e3eb6b91ef9 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 565.654766] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Releasing lock "refresh_cache-5ebc4807-907b-4254-8496-92f1bbb7b761" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.655477] env[62600]: DEBUG nova.compute.manager [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 565.655477] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 565.656305] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d406cff0-9a23-464c-af47-43708ad31607 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.669710] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-812403f1-66a3-4e16-a8a6-f5d0e160b0e1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.697272] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5ebc4807-907b-4254-8496-92f1bbb7b761 could not be found. [ 565.697718] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 565.698154] env[62600]: INFO nova.compute.manager [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Took 0.04 seconds to destroy the instance on the hypervisor. [ 565.699828] env[62600]: DEBUG oslo.service.loopingcall [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 565.699828] env[62600]: DEBUG nova.compute.manager [-] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 565.699828] env[62600]: DEBUG nova.network.neutron [-] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 565.734542] env[62600]: DEBUG nova.virt.hardware [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 565.734801] env[62600]: DEBUG nova.virt.hardware [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 565.734957] env[62600]: DEBUG nova.virt.hardware [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 565.735166] env[62600]: DEBUG nova.virt.hardware [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 565.735320] env[62600]: DEBUG nova.virt.hardware [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 565.735586] env[62600]: DEBUG nova.virt.hardware [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 565.736016] env[62600]: DEBUG nova.virt.hardware [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 565.736278] env[62600]: DEBUG nova.virt.hardware [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 565.736972] env[62600]: DEBUG nova.virt.hardware [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 565.737173] env[62600]: DEBUG nova.virt.hardware [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 565.737355] env[62600]: DEBUG nova.virt.hardware [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 565.738384] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b67cc4-e43d-49fd-9e03-e0ff544b7b34 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.741807] env[62600]: DEBUG nova.network.neutron [-] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 565.753500] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4123b426-1d1c-48f1-a85b-5c27df7133ed {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.774345] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Instance VIF info [] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 565.782696] env[62600]: DEBUG oslo.service.loopingcall [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 565.786494] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 565.786494] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a3b07a51-93e9-440f-b9f7-dc90ae360cae {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.800819] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 565.800819] env[62600]: value = "task-1222192" [ 565.800819] env[62600]: _type = "Task" [ 565.800819] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.809912] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222192, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.027125] env[62600]: DEBUG nova.scheduler.client.report [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 566.099553] env[62600]: DEBUG nova.compute.manager [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 566.130135] env[62600]: DEBUG nova.virt.hardware [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 566.130135] env[62600]: DEBUG nova.virt.hardware [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 566.130135] env[62600]: DEBUG nova.virt.hardware [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 566.130297] env[62600]: DEBUG nova.virt.hardware [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 566.130297] env[62600]: DEBUG nova.virt.hardware [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 566.130297] env[62600]: DEBUG nova.virt.hardware [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 566.130297] env[62600]: DEBUG nova.virt.hardware [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 566.131172] env[62600]: DEBUG nova.virt.hardware [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 566.131611] env[62600]: DEBUG nova.virt.hardware [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 566.132049] env[62600]: DEBUG nova.virt.hardware [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 566.133550] env[62600]: DEBUG nova.virt.hardware [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 566.134043] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b5d1aff-eb1a-46e4-868b-2b5d8bea962c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.147670] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57a398f9-d996-4b53-a40c-0ea999b79cc9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.245827] env[62600]: DEBUG nova.network.neutron [-] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.314367] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222192, 'name': CreateVM_Task, 'duration_secs': 0.301582} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.314367] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 566.315297] env[62600]: DEBUG oslo_vmware.service [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290f6a93-312a-48aa-b5bd-0f33427211ce {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.322050] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.322050] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.322409] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 566.322700] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcdb0628-a8f3-4d0b-a1e5-f3d44e0367ff {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.331821] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 566.331821] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525db120-ceac-2ce1-d2d7-fe7e8542507d" [ 566.331821] env[62600]: _type = "Task" [ 566.331821] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.340392] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525db120-ceac-2ce1-d2d7-fe7e8542507d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.535025] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.463s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.535025] env[62600]: DEBUG nova.compute.manager [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 566.541793] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.331s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.545341] env[62600]: INFO nova.compute.claims [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 566.755018] env[62600]: INFO nova.compute.manager [-] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Took 1.05 seconds to deallocate network for instance. [ 566.758773] env[62600]: DEBUG nova.compute.claims [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 566.760360] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.847634] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.847874] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 566.848106] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.848272] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.848462] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 566.848716] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-da3df6cd-4596-43c1-ae59-0d288781dc3d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.873786] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 566.873786] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 566.874273] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d94484d-46ae-4403-823a-d62258ce6518 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.883514] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f911f20-1938-4aea-89fc-3b0732719ac6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.892398] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 566.892398] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525b4c93-5070-3592-8e55-b09774ce7f89" [ 566.892398] env[62600]: _type = "Task" [ 566.892398] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.900857] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525b4c93-5070-3592-8e55-b09774ce7f89, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.921289] env[62600]: DEBUG oslo_concurrency.lockutils [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquiring lock "222a94b5-7841-490e-9b27-d3f703e6794f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.923025] env[62600]: DEBUG oslo_concurrency.lockutils [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Lock "222a94b5-7841-490e-9b27-d3f703e6794f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.044208] env[62600]: DEBUG nova.compute.utils [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 567.051131] env[62600]: DEBUG nova.compute.manager [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 567.051131] env[62600]: DEBUG nova.network.neutron [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 567.170453] env[62600]: DEBUG nova.policy [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '46e18992b34d412e8e3e8205d1d18265', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '734eda84fbe4493ba1f3db243a2bdb09', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 567.240269] env[62600]: DEBUG nova.compute.manager [req-a8687575-aef8-40ec-b705-8c74b03f10a8 req-10852af5-8269-4ce4-add6-657526377c06 service nova] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Received event network-changed-a10a7f80-347e-49c9-aae5-76ca002d1a38 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 567.240269] env[62600]: DEBUG nova.compute.manager [req-a8687575-aef8-40ec-b705-8c74b03f10a8 req-10852af5-8269-4ce4-add6-657526377c06 service nova] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Refreshing instance network info cache due to event network-changed-a10a7f80-347e-49c9-aae5-76ca002d1a38. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 567.240269] env[62600]: DEBUG oslo_concurrency.lockutils [req-a8687575-aef8-40ec-b705-8c74b03f10a8 req-10852af5-8269-4ce4-add6-657526377c06 service nova] Acquiring lock "refresh_cache-5ebc4807-907b-4254-8496-92f1bbb7b761" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.240269] env[62600]: DEBUG oslo_concurrency.lockutils [req-a8687575-aef8-40ec-b705-8c74b03f10a8 req-10852af5-8269-4ce4-add6-657526377c06 service nova] Acquired lock "refresh_cache-5ebc4807-907b-4254-8496-92f1bbb7b761" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.240269] env[62600]: DEBUG nova.network.neutron [req-a8687575-aef8-40ec-b705-8c74b03f10a8 req-10852af5-8269-4ce4-add6-657526377c06 service nova] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Refreshing network info cache for port a10a7f80-347e-49c9-aae5-76ca002d1a38 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 567.404803] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Preparing fetch location {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 567.406449] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Creating directory with path [datastore1] vmware_temp/5a76a86b-5241-4c27-9511-4f95542befc0/e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 567.406695] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-05711dfc-a174-47f8-a414-9431b89b3c49 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.426898] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Created directory with path [datastore1] vmware_temp/5a76a86b-5241-4c27-9511-4f95542befc0/e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 567.427121] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Fetch image to [datastore1] vmware_temp/5a76a86b-5241-4c27-9511-4f95542befc0/e7bd1d77-bfff-4684-9545-f6a1c69efe58/tmp-sparse.vmdk {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 567.427361] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Downloading image file data e7bd1d77-bfff-4684-9545-f6a1c69efe58 to [datastore1] vmware_temp/5a76a86b-5241-4c27-9511-4f95542befc0/e7bd1d77-bfff-4684-9545-f6a1c69efe58/tmp-sparse.vmdk on the data store datastore1 {{(pid=62600) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 567.428088] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3da9c5d3-7bf8-415b-b7e4-fa6a950e595e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.435815] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392401d6-757e-47f4-95d0-8e846324d427 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.445057] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cb81f3c-0d37-4b2b-81a3-679f85df26b9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.479901] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-900867e7-1c4e-4669-adc5-a8f885eed55c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.491648] env[62600]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-adb64d83-0ca5-4088-a69a-0b4b9b3d2668 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.549473] env[62600]: DEBUG nova.compute.manager [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 567.583476] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Downloading image file data e7bd1d77-bfff-4684-9545-f6a1c69efe58 to the data store datastore1 {{(pid=62600) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 567.662436] env[62600]: DEBUG oslo_vmware.rw_handles [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/5a76a86b-5241-4c27-9511-4f95542befc0/e7bd1d77-bfff-4684-9545-f6a1c69efe58/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62600) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 567.776750] env[62600]: DEBUG nova.network.neutron [req-a8687575-aef8-40ec-b705-8c74b03f10a8 req-10852af5-8269-4ce4-add6-657526377c06 service nova] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.780519] env[62600]: DEBUG nova.network.neutron [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Successfully created port: 60d470c3-2cd9-4861-8309-78911c4b1d06 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 567.988751] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-859910fb-3e15-41b2-bc9b-439801355776 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.996526] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08adae2c-a440-408c-9da2-291c79fc800a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.029279] env[62600]: ERROR nova.compute.manager [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 987155ef-8896-4776-83a8-4ac68b0e1f18, please check neutron logs for more information. [ 568.029279] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 568.029279] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.029279] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 568.029279] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 568.029279] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 568.029279] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 568.029279] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 568.029279] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.029279] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 568.029279] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.029279] env[62600]: ERROR nova.compute.manager raise self.value [ 568.029279] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 568.029279] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 568.029279] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.029279] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 568.029817] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.029817] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 568.029817] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 987155ef-8896-4776-83a8-4ac68b0e1f18, please check neutron logs for more information. [ 568.029817] env[62600]: ERROR nova.compute.manager [ 568.029817] env[62600]: Traceback (most recent call last): [ 568.029817] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 568.029817] env[62600]: listener.cb(fileno) [ 568.029817] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.029817] env[62600]: result = function(*args, **kwargs) [ 568.029817] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.029817] env[62600]: return func(*args, **kwargs) [ 568.029817] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.029817] env[62600]: raise e [ 568.029817] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.029817] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 568.029817] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 568.029817] env[62600]: created_port_ids = self._update_ports_for_instance( [ 568.029817] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 568.029817] env[62600]: with excutils.save_and_reraise_exception(): [ 568.029817] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.029817] env[62600]: self.force_reraise() [ 568.029817] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.029817] env[62600]: raise self.value [ 568.029817] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 568.029817] env[62600]: updated_port = self._update_port( [ 568.029817] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.029817] env[62600]: _ensure_no_port_binding_failure(port) [ 568.029817] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.029817] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 568.030621] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 987155ef-8896-4776-83a8-4ac68b0e1f18, please check neutron logs for more information. [ 568.030621] env[62600]: Removing descriptor: 16 [ 568.030621] env[62600]: DEBUG nova.network.neutron [req-a8687575-aef8-40ec-b705-8c74b03f10a8 req-10852af5-8269-4ce4-add6-657526377c06 service nova] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.035588] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-923c2a7b-6107-47c9-92c2-572800f8643c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.038918] env[62600]: ERROR nova.compute.manager [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 987155ef-8896-4776-83a8-4ac68b0e1f18, please check neutron logs for more information. [ 568.038918] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Traceback (most recent call last): [ 568.038918] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 568.038918] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] yield resources [ 568.038918] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 568.038918] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] self.driver.spawn(context, instance, image_meta, [ 568.038918] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 568.038918] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] self._vmops.spawn(context, instance, image_meta, injected_files, [ 568.038918] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 568.038918] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] vm_ref = self.build_virtual_machine(instance, [ 568.038918] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 568.039525] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] vif_infos = vmwarevif.get_vif_info(self._session, [ 568.039525] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 568.039525] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] for vif in network_info: [ 568.039525] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 568.039525] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] return self._sync_wrapper(fn, *args, **kwargs) [ 568.039525] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 568.039525] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] self.wait() [ 568.039525] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 568.039525] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] self[:] = self._gt.wait() [ 568.039525] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 568.039525] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] return self._exit_event.wait() [ 568.039525] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 568.039525] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] result = hub.switch() [ 568.040456] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 568.040456] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] return self.greenlet.switch() [ 568.040456] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.040456] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] result = function(*args, **kwargs) [ 568.040456] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.040456] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] return func(*args, **kwargs) [ 568.040456] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.040456] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] raise e [ 568.040456] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.040456] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] nwinfo = self.network_api.allocate_for_instance( [ 568.040456] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 568.040456] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] created_port_ids = self._update_ports_for_instance( [ 568.040456] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 568.041452] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] with excutils.save_and_reraise_exception(): [ 568.041452] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.041452] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] self.force_reraise() [ 568.041452] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.041452] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] raise self.value [ 568.041452] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 568.041452] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] updated_port = self._update_port( [ 568.041452] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.041452] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] _ensure_no_port_binding_failure(port) [ 568.041452] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.041452] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] raise exception.PortBindingFailed(port_id=port['id']) [ 568.041452] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] nova.exception.PortBindingFailed: Binding failed for port 987155ef-8896-4776-83a8-4ac68b0e1f18, please check neutron logs for more information. [ 568.041452] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] [ 568.042032] env[62600]: INFO nova.compute.manager [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Terminating instance [ 568.044574] env[62600]: DEBUG oslo_concurrency.lockutils [req-a8687575-aef8-40ec-b705-8c74b03f10a8 req-10852af5-8269-4ce4-add6-657526377c06 service nova] Releasing lock "refresh_cache-5ebc4807-907b-4254-8496-92f1bbb7b761" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.044803] env[62600]: DEBUG nova.compute.manager [req-a8687575-aef8-40ec-b705-8c74b03f10a8 req-10852af5-8269-4ce4-add6-657526377c06 service nova] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Received event network-vif-deleted-a10a7f80-347e-49c9-aae5-76ca002d1a38 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 568.045382] env[62600]: DEBUG oslo_concurrency.lockutils [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "refresh_cache-b74bf924-2d0b-4221-9c82-aa5b1ef903ef" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.045540] env[62600]: DEBUG oslo_concurrency.lockutils [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired lock "refresh_cache-b74bf924-2d0b-4221-9c82-aa5b1ef903ef" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.045702] env[62600]: DEBUG nova.network.neutron [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 568.052249] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a117d203-54bf-42f6-94e4-1268f3501c42 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.069981] env[62600]: DEBUG nova.compute.provider_tree [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 568.464633] env[62600]: DEBUG oslo_vmware.rw_handles [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Completed reading data from the image iterator. {{(pid=62600) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 568.464633] env[62600]: DEBUG oslo_vmware.rw_handles [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/5a76a86b-5241-4c27-9511-4f95542befc0/e7bd1d77-bfff-4684-9545-f6a1c69efe58/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62600) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 568.526233] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Downloaded image file data e7bd1d77-bfff-4684-9545-f6a1c69efe58 to vmware_temp/5a76a86b-5241-4c27-9511-4f95542befc0/e7bd1d77-bfff-4684-9545-f6a1c69efe58/tmp-sparse.vmdk on the data store datastore1 {{(pid=62600) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 568.530668] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Caching image {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 568.531128] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Copying Virtual Disk [datastore1] vmware_temp/5a76a86b-5241-4c27-9511-4f95542befc0/e7bd1d77-bfff-4684-9545-f6a1c69efe58/tmp-sparse.vmdk to [datastore1] vmware_temp/5a76a86b-5241-4c27-9511-4f95542befc0/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 568.531355] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3039240a-9b89-4558-a1d1-60f999eeeda0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.541540] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 568.541540] env[62600]: value = "task-1222193" [ 568.541540] env[62600]: _type = "Task" [ 568.541540] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.552596] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222193, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.572784] env[62600]: DEBUG nova.compute.manager [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 568.576676] env[62600]: DEBUG nova.scheduler.client.report [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 568.586898] env[62600]: DEBUG nova.network.neutron [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 568.610736] env[62600]: DEBUG nova.virt.hardware [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 568.610736] env[62600]: DEBUG nova.virt.hardware [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 568.610736] env[62600]: DEBUG nova.virt.hardware [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 568.611028] env[62600]: DEBUG nova.virt.hardware [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 568.611028] env[62600]: DEBUG nova.virt.hardware [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 568.611252] env[62600]: DEBUG nova.virt.hardware [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 568.611535] env[62600]: DEBUG nova.virt.hardware [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 568.611660] env[62600]: DEBUG nova.virt.hardware [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 568.611858] env[62600]: DEBUG nova.virt.hardware [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 568.612072] env[62600]: DEBUG nova.virt.hardware [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 568.612522] env[62600]: DEBUG nova.virt.hardware [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 568.613684] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e1fdd22-2b44-4cd2-98f2-d7cd1c418179 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.623895] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-121b68ae-9666-4d5f-870a-6253846e88f8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.777609] env[62600]: DEBUG nova.network.neutron [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.052667] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222193, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.083271] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.542s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.084383] env[62600]: DEBUG nova.compute.manager [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 569.087028] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.526s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.088483] env[62600]: INFO nova.compute.claims [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 569.281449] env[62600]: DEBUG oslo_concurrency.lockutils [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Releasing lock "refresh_cache-b74bf924-2d0b-4221-9c82-aa5b1ef903ef" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 569.282136] env[62600]: DEBUG nova.compute.manager [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 569.282404] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 569.282698] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6f555f49-ab16-4dc6-8026-da618c5bc348 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.295073] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b24b21f3-63dd-4cf6-825e-062c3bd0a31c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.318751] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b74bf924-2d0b-4221-9c82-aa5b1ef903ef could not be found. [ 569.318998] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 569.319278] env[62600]: INFO nova.compute.manager [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Took 0.04 seconds to destroy the instance on the hypervisor. [ 569.319584] env[62600]: DEBUG oslo.service.loopingcall [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 569.319917] env[62600]: DEBUG nova.compute.manager [-] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 569.320709] env[62600]: DEBUG nova.network.neutron [-] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 569.353127] env[62600]: DEBUG nova.network.neutron [-] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 569.557110] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222193, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.651732} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.557424] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Copied Virtual Disk [datastore1] vmware_temp/5a76a86b-5241-4c27-9511-4f95542befc0/e7bd1d77-bfff-4684-9545-f6a1c69efe58/tmp-sparse.vmdk to [datastore1] vmware_temp/5a76a86b-5241-4c27-9511-4f95542befc0/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 569.557607] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Deleting the datastore file [datastore1] vmware_temp/5a76a86b-5241-4c27-9511-4f95542befc0/e7bd1d77-bfff-4684-9545-f6a1c69efe58/tmp-sparse.vmdk {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 569.557882] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-484f9249-4351-4d6d-8e94-747e5b66ac6a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.564954] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 569.564954] env[62600]: value = "task-1222194" [ 569.564954] env[62600]: _type = "Task" [ 569.564954] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.575911] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222194, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.594961] env[62600]: DEBUG nova.compute.utils [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 569.598071] env[62600]: DEBUG nova.compute.manager [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 569.598405] env[62600]: DEBUG nova.network.neutron [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 569.676509] env[62600]: DEBUG nova.compute.manager [req-c5c8da1f-c433-40f7-8c31-be75df4949df req-71138292-e53d-4779-9d66-81fd712a1cd7 service nova] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Received event network-changed-987155ef-8896-4776-83a8-4ac68b0e1f18 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 569.676954] env[62600]: DEBUG nova.compute.manager [req-c5c8da1f-c433-40f7-8c31-be75df4949df req-71138292-e53d-4779-9d66-81fd712a1cd7 service nova] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Refreshing instance network info cache due to event network-changed-987155ef-8896-4776-83a8-4ac68b0e1f18. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 569.676954] env[62600]: DEBUG oslo_concurrency.lockutils [req-c5c8da1f-c433-40f7-8c31-be75df4949df req-71138292-e53d-4779-9d66-81fd712a1cd7 service nova] Acquiring lock "refresh_cache-b74bf924-2d0b-4221-9c82-aa5b1ef903ef" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.677088] env[62600]: DEBUG oslo_concurrency.lockutils [req-c5c8da1f-c433-40f7-8c31-be75df4949df req-71138292-e53d-4779-9d66-81fd712a1cd7 service nova] Acquired lock "refresh_cache-b74bf924-2d0b-4221-9c82-aa5b1ef903ef" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.679066] env[62600]: DEBUG nova.network.neutron [req-c5c8da1f-c433-40f7-8c31-be75df4949df req-71138292-e53d-4779-9d66-81fd712a1cd7 service nova] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Refreshing network info cache for port 987155ef-8896-4776-83a8-4ac68b0e1f18 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 569.811927] env[62600]: DEBUG nova.policy [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88b7fd08dd8b488cab7d37388d7874fc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ef9dbdf45ae4b91afc7c83c49d3c171', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 569.852942] env[62600]: DEBUG nova.network.neutron [-] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.859589] env[62600]: ERROR nova.compute.manager [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 60d470c3-2cd9-4861-8309-78911c4b1d06, please check neutron logs for more information. [ 569.859589] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 569.859589] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.859589] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 569.859589] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 569.859589] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 569.859589] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 569.859589] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 569.859589] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.859589] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 569.859589] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.859589] env[62600]: ERROR nova.compute.manager raise self.value [ 569.859589] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 569.859589] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 569.859589] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.859589] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 569.860305] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.860305] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 569.860305] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 60d470c3-2cd9-4861-8309-78911c4b1d06, please check neutron logs for more information. [ 569.860305] env[62600]: ERROR nova.compute.manager [ 569.860305] env[62600]: Traceback (most recent call last): [ 569.860305] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 569.860305] env[62600]: listener.cb(fileno) [ 569.860305] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.860305] env[62600]: result = function(*args, **kwargs) [ 569.860305] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 569.860305] env[62600]: return func(*args, **kwargs) [ 569.860305] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 569.860305] env[62600]: raise e [ 569.860305] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.860305] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 569.860305] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 569.860305] env[62600]: created_port_ids = self._update_ports_for_instance( [ 569.860305] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 569.860305] env[62600]: with excutils.save_and_reraise_exception(): [ 569.860305] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.860305] env[62600]: self.force_reraise() [ 569.860305] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.860305] env[62600]: raise self.value [ 569.860305] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 569.860305] env[62600]: updated_port = self._update_port( [ 569.860305] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.860305] env[62600]: _ensure_no_port_binding_failure(port) [ 569.860305] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.860305] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 569.861253] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 60d470c3-2cd9-4861-8309-78911c4b1d06, please check neutron logs for more information. [ 569.861253] env[62600]: Removing descriptor: 15 [ 569.861253] env[62600]: ERROR nova.compute.manager [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 60d470c3-2cd9-4861-8309-78911c4b1d06, please check neutron logs for more information. [ 569.861253] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Traceback (most recent call last): [ 569.861253] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 569.861253] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] yield resources [ 569.861253] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 569.861253] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] self.driver.spawn(context, instance, image_meta, [ 569.861253] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 569.861253] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 569.861253] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 569.861253] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] vm_ref = self.build_virtual_machine(instance, [ 569.861969] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 569.861969] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] vif_infos = vmwarevif.get_vif_info(self._session, [ 569.861969] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 569.861969] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] for vif in network_info: [ 569.861969] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 569.861969] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] return self._sync_wrapper(fn, *args, **kwargs) [ 569.861969] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 569.861969] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] self.wait() [ 569.861969] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 569.861969] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] self[:] = self._gt.wait() [ 569.861969] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 569.861969] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] return self._exit_event.wait() [ 569.861969] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 569.862329] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] result = hub.switch() [ 569.862329] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 569.862329] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] return self.greenlet.switch() [ 569.862329] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.862329] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] result = function(*args, **kwargs) [ 569.862329] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 569.862329] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] return func(*args, **kwargs) [ 569.862329] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 569.862329] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] raise e [ 569.862329] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.862329] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] nwinfo = self.network_api.allocate_for_instance( [ 569.862329] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 569.862329] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] created_port_ids = self._update_ports_for_instance( [ 569.862931] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 569.862931] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] with excutils.save_and_reraise_exception(): [ 569.862931] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.862931] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] self.force_reraise() [ 569.862931] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.862931] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] raise self.value [ 569.862931] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 569.862931] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] updated_port = self._update_port( [ 569.862931] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.862931] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] _ensure_no_port_binding_failure(port) [ 569.862931] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.862931] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] raise exception.PortBindingFailed(port_id=port['id']) [ 569.863294] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] nova.exception.PortBindingFailed: Binding failed for port 60d470c3-2cd9-4861-8309-78911c4b1d06, please check neutron logs for more information. [ 569.863294] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] [ 569.863294] env[62600]: INFO nova.compute.manager [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Terminating instance [ 569.864457] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "refresh_cache-a664416f-f9f6-4cdd-93b0-8dd63181562e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.864617] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquired lock "refresh_cache-a664416f-f9f6-4cdd-93b0-8dd63181562e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.864778] env[62600]: DEBUG nova.network.neutron [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 570.081219] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222194, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023671} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 570.081498] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 570.081720] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Moving file from [datastore1] vmware_temp/5a76a86b-5241-4c27-9511-4f95542befc0/e7bd1d77-bfff-4684-9545-f6a1c69efe58 to [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58. {{(pid=62600) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 570.082587] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-5d3ff994-4b0a-429c-9db7-5fd9c590da88 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.093106] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 570.093106] env[62600]: value = "task-1222195" [ 570.093106] env[62600]: _type = "Task" [ 570.093106] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.102546] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222195, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.109355] env[62600]: DEBUG nova.compute.manager [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 570.205287] env[62600]: DEBUG nova.network.neutron [req-c5c8da1f-c433-40f7-8c31-be75df4949df req-71138292-e53d-4779-9d66-81fd712a1cd7 service nova] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 570.364213] env[62600]: DEBUG nova.network.neutron [req-c5c8da1f-c433-40f7-8c31-be75df4949df req-71138292-e53d-4779-9d66-81fd712a1cd7 service nova] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.365966] env[62600]: INFO nova.compute.manager [-] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Took 1.05 seconds to deallocate network for instance. [ 570.374833] env[62600]: DEBUG nova.compute.claims [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 570.374833] env[62600]: DEBUG oslo_concurrency.lockutils [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.391945] env[62600]: DEBUG nova.network.neutron [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 570.537979] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74fe70a7-3f1b-4bfb-b28f-d0fc227039e0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.543911] env[62600]: DEBUG nova.network.neutron [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.552721] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01a3431c-ea43-4df3-a543-7419a7798f57 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.599587] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17d2c3eb-1398-40e8-8c15-1ff3ac0cc09e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.607593] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222195, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.025075} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 570.610582] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] File moved {{(pid=62600) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 570.611415] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Cleaning up location [datastore1] vmware_temp/5a76a86b-5241-4c27-9511-4f95542befc0 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 570.611924] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Deleting the datastore file [datastore1] vmware_temp/5a76a86b-5241-4c27-9511-4f95542befc0 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 570.612628] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db56b964-f6d0-4eb2-a80e-7ac3c485f52a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.616253] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac976fe2-9fe6-4a7d-86de-f722699acf6d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.637119] env[62600]: DEBUG nova.compute.provider_tree [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 570.640511] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 570.640511] env[62600]: value = "task-1222196" [ 570.640511] env[62600]: _type = "Task" [ 570.640511] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.652233] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222196, 'name': DeleteDatastoreFile_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.711512] env[62600]: DEBUG nova.network.neutron [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Successfully created port: 9c6a925d-932e-4c78-9c8b-aff2cce4382b {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 570.871845] env[62600]: DEBUG oslo_concurrency.lockutils [req-c5c8da1f-c433-40f7-8c31-be75df4949df req-71138292-e53d-4779-9d66-81fd712a1cd7 service nova] Releasing lock "refresh_cache-b74bf924-2d0b-4221-9c82-aa5b1ef903ef" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.873065] env[62600]: DEBUG nova.compute.manager [req-c5c8da1f-c433-40f7-8c31-be75df4949df req-71138292-e53d-4779-9d66-81fd712a1cd7 service nova] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Received event network-vif-deleted-987155ef-8896-4776-83a8-4ac68b0e1f18 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 571.050536] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Releasing lock "refresh_cache-a664416f-f9f6-4cdd-93b0-8dd63181562e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.050725] env[62600]: DEBUG nova.compute.manager [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 571.050916] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 571.051301] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a0a95402-1f37-4188-9275-bb8e9311e81f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.066754] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89719bf-d6ce-453c-9f85-24345e882457 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.090132] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a664416f-f9f6-4cdd-93b0-8dd63181562e could not be found. [ 571.090345] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 571.090535] env[62600]: INFO nova.compute.manager [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 571.090966] env[62600]: DEBUG oslo.service.loopingcall [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 571.091139] env[62600]: DEBUG nova.compute.manager [-] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 571.091172] env[62600]: DEBUG nova.network.neutron [-] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 571.106983] env[62600]: DEBUG nova.network.neutron [-] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 571.140890] env[62600]: DEBUG nova.compute.manager [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 571.143686] env[62600]: DEBUG nova.scheduler.client.report [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 571.160759] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222196, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024807} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.161498] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 571.163909] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75e84f1d-c813-4402-bf83-f9e75fdef28d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.172362] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 571.172362] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c2aaf9-b303-9cd3-9539-991849844074" [ 571.172362] env[62600]: _type = "Task" [ 571.172362] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.178571] env[62600]: DEBUG nova.virt.hardware [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 571.178796] env[62600]: DEBUG nova.virt.hardware [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 571.178946] env[62600]: DEBUG nova.virt.hardware [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 571.179136] env[62600]: DEBUG nova.virt.hardware [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 571.182038] env[62600]: DEBUG nova.virt.hardware [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 571.182038] env[62600]: DEBUG nova.virt.hardware [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 571.182038] env[62600]: DEBUG nova.virt.hardware [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 571.182038] env[62600]: DEBUG nova.virt.hardware [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 571.182038] env[62600]: DEBUG nova.virt.hardware [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 571.182334] env[62600]: DEBUG nova.virt.hardware [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 571.182334] env[62600]: DEBUG nova.virt.hardware [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 571.182334] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f30f22f-02bd-4610-9823-5c2cdd9eafc6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.190640] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c2aaf9-b303-9cd3-9539-991849844074, 'name': SearchDatastore_Task, 'duration_secs': 0.008893} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.192801] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.193155] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 98cbcc28-54e0-40de-9feb-6da917806423/98cbcc28-54e0-40de-9feb-6da917806423.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 571.193722] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bbe800e5-38bf-4c7e-b189-99c8e8d62982 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.196752] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb8d4c3-30b0-4653-a52a-6b1088ee33b5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.215013] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 571.215013] env[62600]: value = "task-1222197" [ 571.215013] env[62600]: _type = "Task" [ 571.215013] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.612732] env[62600]: DEBUG nova.network.neutron [-] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.655015] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.568s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.658506] env[62600]: DEBUG nova.compute.manager [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 571.660869] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 14.913s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.664802] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.664802] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62600) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 571.664802] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.626s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.669315] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212515cc-ec3a-40a7-8c62-42dde8da3e27 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.684223] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c602916-5556-458d-8ddd-a946e9bb3324 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.702926] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85734a01-ff3b-4d92-8824-bb33c11dde89 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.711022] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-990c1505-ae50-49bb-93ec-8a0528274bae {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.759182] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181479MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=62600) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 571.759354] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.759872] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222197, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.498523} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.760393] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 98cbcc28-54e0-40de-9feb-6da917806423/98cbcc28-54e0-40de-9feb-6da917806423.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 571.760646] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 571.760976] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2d74f8ea-9c32-40be-8cde-7a93c2218133 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.769309] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 571.769309] env[62600]: value = "task-1222198" [ 571.769309] env[62600]: _type = "Task" [ 571.769309] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.782768] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222198, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.119841] env[62600]: INFO nova.compute.manager [-] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Took 1.03 seconds to deallocate network for instance. [ 572.121212] env[62600]: DEBUG nova.compute.claims [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 572.121836] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.163864] env[62600]: DEBUG nova.compute.utils [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 572.164641] env[62600]: DEBUG nova.compute.manager [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 572.164795] env[62600]: DEBUG nova.network.neutron [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 572.263391] env[62600]: DEBUG nova.policy [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7627870c06e14be4aedbd4ee9bee374f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8447a3f29a8347a0b3be942f5bdbc126', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 572.280310] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222198, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.185204} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 572.280604] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 572.281561] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e69c14f7-8012-48d6-b1ad-52859fdde950 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.310618] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Reconfiguring VM instance instance-00000001 to attach disk [datastore1] 98cbcc28-54e0-40de-9feb-6da917806423/98cbcc28-54e0-40de-9feb-6da917806423.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 572.314490] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aa7fdff6-0785-4b33-80c3-3f5ef33f0bca {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.336419] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 572.336419] env[62600]: value = "task-1222199" [ 572.336419] env[62600]: _type = "Task" [ 572.336419] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 572.345335] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222199, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.570072] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb7db06c-62fb-4e77-bc5d-b209ea275a45 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.579971] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03e4b034-f257-4ea8-ac0e-ef90c4857e06 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.615234] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b65cfdd5-847f-4427-a7a2-1b3a4119604a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.623382] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6c21bfa-b2e6-419d-aa1e-d0b91388a4e9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.637736] env[62600]: DEBUG nova.compute.provider_tree [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 572.669572] env[62600]: DEBUG nova.compute.manager [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 572.681198] env[62600]: DEBUG nova.network.neutron [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Successfully created port: 25a47394-9295-4585-8463-7d56a4fba500 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 572.847104] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222199, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.971137] env[62600]: DEBUG nova.compute.manager [req-6337a9c4-555a-43e5-a7b3-321d7111ebce req-26a907c3-8c3f-4b3d-a15a-89e962d66fd5 service nova] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Received event network-changed-60d470c3-2cd9-4861-8309-78911c4b1d06 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 572.971137] env[62600]: DEBUG nova.compute.manager [req-6337a9c4-555a-43e5-a7b3-321d7111ebce req-26a907c3-8c3f-4b3d-a15a-89e962d66fd5 service nova] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Refreshing instance network info cache due to event network-changed-60d470c3-2cd9-4861-8309-78911c4b1d06. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 572.971137] env[62600]: DEBUG oslo_concurrency.lockutils [req-6337a9c4-555a-43e5-a7b3-321d7111ebce req-26a907c3-8c3f-4b3d-a15a-89e962d66fd5 service nova] Acquiring lock "refresh_cache-a664416f-f9f6-4cdd-93b0-8dd63181562e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.971137] env[62600]: DEBUG oslo_concurrency.lockutils [req-6337a9c4-555a-43e5-a7b3-321d7111ebce req-26a907c3-8c3f-4b3d-a15a-89e962d66fd5 service nova] Acquired lock "refresh_cache-a664416f-f9f6-4cdd-93b0-8dd63181562e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.971910] env[62600]: DEBUG nova.network.neutron [req-6337a9c4-555a-43e5-a7b3-321d7111ebce req-26a907c3-8c3f-4b3d-a15a-89e962d66fd5 service nova] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Refreshing network info cache for port 60d470c3-2cd9-4861-8309-78911c4b1d06 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 572.974545] env[62600]: DEBUG nova.network.neutron [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Successfully created port: 6f36181f-bb21-4fe5-b390-fed6112f550d {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 573.141611] env[62600]: DEBUG nova.scheduler.client.report [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 573.220411] env[62600]: ERROR nova.compute.manager [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9c6a925d-932e-4c78-9c8b-aff2cce4382b, please check neutron logs for more information. [ 573.220411] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 573.220411] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.220411] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 573.220411] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.220411] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 573.220411] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.220411] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 573.220411] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.220411] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 573.220411] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.220411] env[62600]: ERROR nova.compute.manager raise self.value [ 573.220411] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.220411] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 573.220411] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.220411] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 573.220873] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.220873] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 573.220873] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9c6a925d-932e-4c78-9c8b-aff2cce4382b, please check neutron logs for more information. [ 573.220873] env[62600]: ERROR nova.compute.manager [ 573.220873] env[62600]: Traceback (most recent call last): [ 573.220873] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 573.220873] env[62600]: listener.cb(fileno) [ 573.220873] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.220873] env[62600]: result = function(*args, **kwargs) [ 573.220873] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.220873] env[62600]: return func(*args, **kwargs) [ 573.220873] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.220873] env[62600]: raise e [ 573.220873] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.220873] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 573.220873] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.220873] env[62600]: created_port_ids = self._update_ports_for_instance( [ 573.220873] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.220873] env[62600]: with excutils.save_and_reraise_exception(): [ 573.220873] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.220873] env[62600]: self.force_reraise() [ 573.220873] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.220873] env[62600]: raise self.value [ 573.220873] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.220873] env[62600]: updated_port = self._update_port( [ 573.220873] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.220873] env[62600]: _ensure_no_port_binding_failure(port) [ 573.220873] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.220873] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 573.222943] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 9c6a925d-932e-4c78-9c8b-aff2cce4382b, please check neutron logs for more information. [ 573.222943] env[62600]: Removing descriptor: 17 [ 573.222943] env[62600]: ERROR nova.compute.manager [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9c6a925d-932e-4c78-9c8b-aff2cce4382b, please check neutron logs for more information. [ 573.222943] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Traceback (most recent call last): [ 573.222943] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 573.222943] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] yield resources [ 573.222943] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 573.222943] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] self.driver.spawn(context, instance, image_meta, [ 573.222943] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 573.222943] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.222943] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.222943] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] vm_ref = self.build_virtual_machine(instance, [ 573.223510] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.223510] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.223510] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.223510] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] for vif in network_info: [ 573.223510] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 573.223510] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] return self._sync_wrapper(fn, *args, **kwargs) [ 573.223510] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 573.223510] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] self.wait() [ 573.223510] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 573.223510] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] self[:] = self._gt.wait() [ 573.223510] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.223510] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] return self._exit_event.wait() [ 573.223510] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 573.223912] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] result = hub.switch() [ 573.223912] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 573.223912] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] return self.greenlet.switch() [ 573.223912] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.223912] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] result = function(*args, **kwargs) [ 573.223912] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.223912] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] return func(*args, **kwargs) [ 573.223912] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.223912] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] raise e [ 573.223912] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.223912] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] nwinfo = self.network_api.allocate_for_instance( [ 573.223912] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.223912] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] created_port_ids = self._update_ports_for_instance( [ 573.224308] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.224308] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] with excutils.save_and_reraise_exception(): [ 573.224308] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.224308] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] self.force_reraise() [ 573.224308] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.224308] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] raise self.value [ 573.224308] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.224308] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] updated_port = self._update_port( [ 573.224308] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.224308] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] _ensure_no_port_binding_failure(port) [ 573.224308] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.224308] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] raise exception.PortBindingFailed(port_id=port['id']) [ 573.224675] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] nova.exception.PortBindingFailed: Binding failed for port 9c6a925d-932e-4c78-9c8b-aff2cce4382b, please check neutron logs for more information. [ 573.224675] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] [ 573.224675] env[62600]: INFO nova.compute.manager [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Terminating instance [ 573.225364] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Acquiring lock "refresh_cache-38e57696-33a8-49d8-b061-ae14155ed280" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.225696] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Acquired lock "refresh_cache-38e57696-33a8-49d8-b061-ae14155ed280" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.226260] env[62600]: DEBUG nova.network.neutron [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 573.236587] env[62600]: DEBUG nova.network.neutron [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Successfully created port: 7d2e3a85-6e06-4588-ac5f-ea7374876243 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 573.351382] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222199, 'name': ReconfigVM_Task, 'duration_secs': 0.837462} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 573.351849] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Reconfigured VM instance instance-00000001 to attach disk [datastore1] 98cbcc28-54e0-40de-9feb-6da917806423/98cbcc28-54e0-40de-9feb-6da917806423.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 573.352638] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ac1d1b46-e006-4474-9c8b-92ca3441f053 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.360205] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 573.360205] env[62600]: value = "task-1222200" [ 573.360205] env[62600]: _type = "Task" [ 573.360205] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.368594] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222200, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.504200] env[62600]: DEBUG nova.network.neutron [req-6337a9c4-555a-43e5-a7b3-321d7111ebce req-26a907c3-8c3f-4b3d-a15a-89e962d66fd5 service nova] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 573.575538] env[62600]: DEBUG oslo_concurrency.lockutils [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Acquiring lock "4455c247-a707-4c3a-9323-0110a1290780" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.575779] env[62600]: DEBUG oslo_concurrency.lockutils [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Lock "4455c247-a707-4c3a-9323-0110a1290780" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.650176] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.986s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.650176] env[62600]: ERROR nova.compute.manager [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1339efad-11b7-403a-9a9c-9f70795493ea, please check neutron logs for more information. [ 573.650176] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Traceback (most recent call last): [ 573.650176] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 573.650176] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] self.driver.spawn(context, instance, image_meta, [ 573.650176] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 573.650176] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.650176] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.650176] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] vm_ref = self.build_virtual_machine(instance, [ 573.650847] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.650847] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.650847] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.650847] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] for vif in network_info: [ 573.650847] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 573.650847] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] return self._sync_wrapper(fn, *args, **kwargs) [ 573.650847] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 573.650847] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] self.wait() [ 573.650847] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 573.650847] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] self[:] = self._gt.wait() [ 573.650847] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.650847] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] return self._exit_event.wait() [ 573.650847] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 573.651379] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] result = hub.switch() [ 573.651379] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 573.651379] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] return self.greenlet.switch() [ 573.651379] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.651379] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] result = function(*args, **kwargs) [ 573.651379] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.651379] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] return func(*args, **kwargs) [ 573.651379] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.651379] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] raise e [ 573.651379] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.651379] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] nwinfo = self.network_api.allocate_for_instance( [ 573.651379] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.651379] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] created_port_ids = self._update_ports_for_instance( [ 573.651918] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.651918] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] with excutils.save_and_reraise_exception(): [ 573.651918] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.651918] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] self.force_reraise() [ 573.651918] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.651918] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] raise self.value [ 573.651918] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.651918] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] updated_port = self._update_port( [ 573.651918] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.651918] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] _ensure_no_port_binding_failure(port) [ 573.651918] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.651918] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] raise exception.PortBindingFailed(port_id=port['id']) [ 573.652454] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] nova.exception.PortBindingFailed: Binding failed for port 1339efad-11b7-403a-9a9c-9f70795493ea, please check neutron logs for more information. [ 573.652454] env[62600]: ERROR nova.compute.manager [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] [ 573.652454] env[62600]: DEBUG nova.compute.utils [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Binding failed for port 1339efad-11b7-403a-9a9c-9f70795493ea, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 573.655476] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.807s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.666859] env[62600]: DEBUG nova.compute.manager [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Build of instance da07d0e2-98bc-451c-955e-db5eaeee3c82 was re-scheduled: Binding failed for port 1339efad-11b7-403a-9a9c-9f70795493ea, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 573.668079] env[62600]: DEBUG nova.compute.manager [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 573.668590] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Acquiring lock "refresh_cache-da07d0e2-98bc-451c-955e-db5eaeee3c82" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.668865] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Acquired lock "refresh_cache-da07d0e2-98bc-451c-955e-db5eaeee3c82" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.669227] env[62600]: DEBUG nova.network.neutron [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 573.680886] env[62600]: DEBUG nova.compute.manager [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 573.684146] env[62600]: DEBUG nova.network.neutron [req-6337a9c4-555a-43e5-a7b3-321d7111ebce req-26a907c3-8c3f-4b3d-a15a-89e962d66fd5 service nova] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.715449] env[62600]: DEBUG nova.virt.hardware [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 573.715734] env[62600]: DEBUG nova.virt.hardware [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 573.715893] env[62600]: DEBUG nova.virt.hardware [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 573.716195] env[62600]: DEBUG nova.virt.hardware [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 573.716449] env[62600]: DEBUG nova.virt.hardware [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 573.716865] env[62600]: DEBUG nova.virt.hardware [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 573.717101] env[62600]: DEBUG nova.virt.hardware [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 573.717283] env[62600]: DEBUG nova.virt.hardware [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 573.717436] env[62600]: DEBUG nova.virt.hardware [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 573.717806] env[62600]: DEBUG nova.virt.hardware [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 573.718038] env[62600]: DEBUG nova.virt.hardware [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 573.719909] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd14282b-d5ad-4b45-853e-1cc402bc082e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.731347] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c46b0b4d-8982-48f6-971e-9d9ce7d9d9f0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.760815] env[62600]: DEBUG nova.network.neutron [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 573.854925] env[62600]: DEBUG nova.network.neutron [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.870901] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222200, 'name': Rename_Task, 'duration_secs': 0.127505} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 573.871189] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 573.871422] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d0958332-9c54-41c1-b1d7-ff36412129b1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.877479] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 573.877479] env[62600]: value = "task-1222201" [ 573.877479] env[62600]: _type = "Task" [ 573.877479] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.885436] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222201, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.189681] env[62600]: DEBUG oslo_concurrency.lockutils [req-6337a9c4-555a-43e5-a7b3-321d7111ebce req-26a907c3-8c3f-4b3d-a15a-89e962d66fd5 service nova] Releasing lock "refresh_cache-a664416f-f9f6-4cdd-93b0-8dd63181562e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.189979] env[62600]: DEBUG nova.compute.manager [req-6337a9c4-555a-43e5-a7b3-321d7111ebce req-26a907c3-8c3f-4b3d-a15a-89e962d66fd5 service nova] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Received event network-vif-deleted-60d470c3-2cd9-4861-8309-78911c4b1d06 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 574.339328] env[62600]: DEBUG nova.network.neutron [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 574.358429] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Releasing lock "refresh_cache-38e57696-33a8-49d8-b061-ae14155ed280" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.358878] env[62600]: DEBUG nova.compute.manager [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 574.359596] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 574.362082] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c1943723-d5ce-45e6-9ce7-820358bda765 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.374422] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35061ed8-0856-4399-a74c-8b16c253e048 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.387180] env[62600]: ERROR nova.compute.manager [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 25a47394-9295-4585-8463-7d56a4fba500, please check neutron logs for more information. [ 574.387180] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 574.387180] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 574.387180] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 574.387180] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 574.387180] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 574.387180] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 574.387180] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 574.387180] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.387180] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 574.387180] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.387180] env[62600]: ERROR nova.compute.manager raise self.value [ 574.387180] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 574.387180] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 574.387180] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.387180] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 574.387688] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.387688] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 574.387688] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 25a47394-9295-4585-8463-7d56a4fba500, please check neutron logs for more information. [ 574.387688] env[62600]: ERROR nova.compute.manager [ 574.387688] env[62600]: Traceback (most recent call last): [ 574.387688] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 574.387688] env[62600]: listener.cb(fileno) [ 574.387688] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 574.387688] env[62600]: result = function(*args, **kwargs) [ 574.387688] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 574.387688] env[62600]: return func(*args, **kwargs) [ 574.387688] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 574.387688] env[62600]: raise e [ 574.387688] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 574.387688] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 574.387688] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 574.387688] env[62600]: created_port_ids = self._update_ports_for_instance( [ 574.387688] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 574.387688] env[62600]: with excutils.save_and_reraise_exception(): [ 574.387688] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.387688] env[62600]: self.force_reraise() [ 574.387688] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.387688] env[62600]: raise self.value [ 574.387688] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 574.387688] env[62600]: updated_port = self._update_port( [ 574.387688] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.387688] env[62600]: _ensure_no_port_binding_failure(port) [ 574.387688] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.387688] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 574.388484] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 25a47394-9295-4585-8463-7d56a4fba500, please check neutron logs for more information. [ 574.388484] env[62600]: Removing descriptor: 15 [ 574.390637] env[62600]: ERROR nova.compute.manager [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 25a47394-9295-4585-8463-7d56a4fba500, please check neutron logs for more information. [ 574.390637] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Traceback (most recent call last): [ 574.390637] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 574.390637] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] yield resources [ 574.390637] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 574.390637] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] self.driver.spawn(context, instance, image_meta, [ 574.390637] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 574.390637] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 574.390637] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 574.390637] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] vm_ref = self.build_virtual_machine(instance, [ 574.390637] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 574.390991] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] vif_infos = vmwarevif.get_vif_info(self._session, [ 574.390991] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 574.390991] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] for vif in network_info: [ 574.390991] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 574.390991] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] return self._sync_wrapper(fn, *args, **kwargs) [ 574.390991] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 574.390991] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] self.wait() [ 574.390991] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 574.390991] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] self[:] = self._gt.wait() [ 574.390991] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 574.390991] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] return self._exit_event.wait() [ 574.390991] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 574.390991] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] result = hub.switch() [ 574.391348] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 574.391348] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] return self.greenlet.switch() [ 574.391348] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 574.391348] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] result = function(*args, **kwargs) [ 574.391348] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 574.391348] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] return func(*args, **kwargs) [ 574.391348] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 574.391348] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] raise e [ 574.391348] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 574.391348] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] nwinfo = self.network_api.allocate_for_instance( [ 574.391348] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 574.391348] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] created_port_ids = self._update_ports_for_instance( [ 574.391348] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 574.391758] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] with excutils.save_and_reraise_exception(): [ 574.391758] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.391758] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] self.force_reraise() [ 574.391758] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.391758] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] raise self.value [ 574.391758] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 574.391758] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] updated_port = self._update_port( [ 574.391758] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.391758] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] _ensure_no_port_binding_failure(port) [ 574.391758] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.391758] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] raise exception.PortBindingFailed(port_id=port['id']) [ 574.391758] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] nova.exception.PortBindingFailed: Binding failed for port 25a47394-9295-4585-8463-7d56a4fba500, please check neutron logs for more information. [ 574.391758] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] [ 574.392216] env[62600]: INFO nova.compute.manager [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Terminating instance [ 574.396130] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Acquiring lock "refresh_cache-c6cf703e-5b00-44b3-ae8b-b655105a073b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.396703] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Acquired lock "refresh_cache-c6cf703e-5b00-44b3-ae8b-b655105a073b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.396703] env[62600]: DEBUG nova.network.neutron [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 574.402343] env[62600]: DEBUG oslo_vmware.api [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222201, 'name': PowerOnVM_Task, 'duration_secs': 0.425358} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.407141] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 574.407353] env[62600]: DEBUG nova.compute.manager [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 574.407914] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 38e57696-33a8-49d8-b061-ae14155ed280 could not be found. [ 574.408125] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 574.408307] env[62600]: INFO nova.compute.manager [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Took 0.05 seconds to destroy the instance on the hypervisor. [ 574.408537] env[62600]: DEBUG oslo.service.loopingcall [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 574.410194] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beff21b9-28a0-40aa-a3cd-11111d24bd09 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.417698] env[62600]: DEBUG nova.compute.manager [-] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.420163] env[62600]: DEBUG nova.network.neutron [-] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 574.453227] env[62600]: DEBUG nova.network.neutron [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.458548] env[62600]: DEBUG nova.network.neutron [-] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 574.464338] env[62600]: DEBUG nova.compute.manager [req-17159533-2609-4ce6-82df-c27f60bef16f req-582ec2d9-976f-4639-9972-3397d4a3e9cb service nova] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Received event network-changed-9c6a925d-932e-4c78-9c8b-aff2cce4382b {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 574.464338] env[62600]: DEBUG nova.compute.manager [req-17159533-2609-4ce6-82df-c27f60bef16f req-582ec2d9-976f-4639-9972-3397d4a3e9cb service nova] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Refreshing instance network info cache due to event network-changed-9c6a925d-932e-4c78-9c8b-aff2cce4382b. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 574.464338] env[62600]: DEBUG oslo_concurrency.lockutils [req-17159533-2609-4ce6-82df-c27f60bef16f req-582ec2d9-976f-4639-9972-3397d4a3e9cb service nova] Acquiring lock "refresh_cache-38e57696-33a8-49d8-b061-ae14155ed280" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.464338] env[62600]: DEBUG oslo_concurrency.lockutils [req-17159533-2609-4ce6-82df-c27f60bef16f req-582ec2d9-976f-4639-9972-3397d4a3e9cb service nova] Acquired lock "refresh_cache-38e57696-33a8-49d8-b061-ae14155ed280" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.464601] env[62600]: DEBUG nova.network.neutron [req-17159533-2609-4ce6-82df-c27f60bef16f req-582ec2d9-976f-4639-9972-3397d4a3e9cb service nova] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Refreshing network info cache for port 9c6a925d-932e-4c78-9c8b-aff2cce4382b {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 574.599341] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7557a00e-81b7-4afe-a206-1744ded1f128 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.608684] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0141a5-0c14-4115-8e2a-1e8d55be3a71 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.640705] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bcb8717-280a-42f4-81cb-c15fd2655afe {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.645814] env[62600]: DEBUG oslo_concurrency.lockutils [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Acquiring lock "09ccf23b-4763-47e1-afad-59e2b06b2d9e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.646046] env[62600]: DEBUG oslo_concurrency.lockutils [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Lock "09ccf23b-4763-47e1-afad-59e2b06b2d9e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.651190] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce516a2d-b288-4ce6-bfa3-b62c132ef35e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.667688] env[62600]: DEBUG nova.compute.provider_tree [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 574.948296] env[62600]: DEBUG nova.network.neutron [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 574.950426] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.960247] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Releasing lock "refresh_cache-da07d0e2-98bc-451c-955e-db5eaeee3c82" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.960507] env[62600]: DEBUG nova.compute.manager [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 574.960680] env[62600]: DEBUG nova.compute.manager [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.960855] env[62600]: DEBUG nova.network.neutron [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 574.962629] env[62600]: DEBUG nova.network.neutron [-] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.992305] env[62600]: DEBUG nova.network.neutron [req-17159533-2609-4ce6-82df-c27f60bef16f req-582ec2d9-976f-4639-9972-3397d4a3e9cb service nova] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 575.001375] env[62600]: DEBUG nova.network.neutron [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 575.078822] env[62600]: DEBUG nova.network.neutron [req-17159533-2609-4ce6-82df-c27f60bef16f req-582ec2d9-976f-4639-9972-3397d4a3e9cb service nova] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.091809] env[62600]: DEBUG nova.network.neutron [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.170459] env[62600]: DEBUG nova.scheduler.client.report [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 575.466147] env[62600]: INFO nova.compute.manager [-] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Took 1.05 seconds to deallocate network for instance. [ 575.470925] env[62600]: DEBUG nova.compute.claims [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 575.471131] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.504404] env[62600]: DEBUG nova.network.neutron [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.581871] env[62600]: DEBUG oslo_concurrency.lockutils [req-17159533-2609-4ce6-82df-c27f60bef16f req-582ec2d9-976f-4639-9972-3397d4a3e9cb service nova] Releasing lock "refresh_cache-38e57696-33a8-49d8-b061-ae14155ed280" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.595272] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Releasing lock "refresh_cache-c6cf703e-5b00-44b3-ae8b-b655105a073b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.596069] env[62600]: DEBUG nova.compute.manager [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 575.596285] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 575.596630] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b8939721-ed47-4fd0-aa3e-a0d4c0145516 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.606707] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80b21556-2a70-4b2e-8113-96a00b235133 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.629446] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c6cf703e-5b00-44b3-ae8b-b655105a073b could not be found. [ 575.629446] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 575.629605] env[62600]: INFO nova.compute.manager [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 575.629849] env[62600]: DEBUG oslo.service.loopingcall [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 575.630214] env[62600]: DEBUG nova.compute.manager [-] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 575.630214] env[62600]: DEBUG nova.network.neutron [-] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 575.683093] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.025s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.683093] env[62600]: ERROR nova.compute.manager [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 92c2ae25-d1b4-47d0-a419-86065fc4d35c, please check neutron logs for more information. [ 575.683093] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Traceback (most recent call last): [ 575.683093] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 575.683093] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] self.driver.spawn(context, instance, image_meta, [ 575.683093] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 575.683093] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] self._vmops.spawn(context, instance, image_meta, injected_files, [ 575.683093] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 575.683093] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] vm_ref = self.build_virtual_machine(instance, [ 575.683853] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 575.683853] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] vif_infos = vmwarevif.get_vif_info(self._session, [ 575.683853] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 575.683853] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] for vif in network_info: [ 575.683853] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 575.683853] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] return self._sync_wrapper(fn, *args, **kwargs) [ 575.683853] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 575.683853] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] self.wait() [ 575.683853] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 575.683853] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] self[:] = self._gt.wait() [ 575.683853] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 575.683853] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] return self._exit_event.wait() [ 575.683853] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 575.684194] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] result = hub.switch() [ 575.684194] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 575.684194] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] return self.greenlet.switch() [ 575.684194] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.684194] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] result = function(*args, **kwargs) [ 575.684194] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 575.684194] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] return func(*args, **kwargs) [ 575.684194] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.684194] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] raise e [ 575.684194] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.684194] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] nwinfo = self.network_api.allocate_for_instance( [ 575.684194] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 575.684194] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] created_port_ids = self._update_ports_for_instance( [ 575.684514] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 575.684514] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] with excutils.save_and_reraise_exception(): [ 575.684514] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.684514] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] self.force_reraise() [ 575.684514] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.684514] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] raise self.value [ 575.684514] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 575.684514] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] updated_port = self._update_port( [ 575.684514] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.684514] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] _ensure_no_port_binding_failure(port) [ 575.684514] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.684514] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] raise exception.PortBindingFailed(port_id=port['id']) [ 575.684810] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] nova.exception.PortBindingFailed: Binding failed for port 92c2ae25-d1b4-47d0-a419-86065fc4d35c, please check neutron logs for more information. [ 575.684810] env[62600]: ERROR nova.compute.manager [instance: b315463b-f8aa-4b16-a5a2-069458f15358] [ 575.684810] env[62600]: DEBUG nova.compute.utils [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Binding failed for port 92c2ae25-d1b4-47d0-a419-86065fc4d35c, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 575.684810] env[62600]: DEBUG oslo_concurrency.lockutils [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.048s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.693256] env[62600]: DEBUG nova.compute.manager [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Build of instance b315463b-f8aa-4b16-a5a2-069458f15358 was re-scheduled: Binding failed for port 92c2ae25-d1b4-47d0-a419-86065fc4d35c, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 575.693256] env[62600]: DEBUG nova.compute.manager [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 575.693256] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Acquiring lock "refresh_cache-b315463b-f8aa-4b16-a5a2-069458f15358" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.693256] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Acquired lock "refresh_cache-b315463b-f8aa-4b16-a5a2-069458f15358" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.693628] env[62600]: DEBUG nova.network.neutron [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 575.697603] env[62600]: DEBUG nova.network.neutron [-] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 575.807115] env[62600]: DEBUG nova.compute.manager [req-a538f857-2b9f-41c8-adb1-dff0c5ebc3d2 req-77846ae3-e3fb-4445-bfd6-e4922d8de0a3 service nova] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Received event network-changed-25a47394-9295-4585-8463-7d56a4fba500 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 575.807326] env[62600]: DEBUG nova.compute.manager [req-a538f857-2b9f-41c8-adb1-dff0c5ebc3d2 req-77846ae3-e3fb-4445-bfd6-e4922d8de0a3 service nova] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Refreshing instance network info cache due to event network-changed-25a47394-9295-4585-8463-7d56a4fba500. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 575.807595] env[62600]: DEBUG oslo_concurrency.lockutils [req-a538f857-2b9f-41c8-adb1-dff0c5ebc3d2 req-77846ae3-e3fb-4445-bfd6-e4922d8de0a3 service nova] Acquiring lock "refresh_cache-c6cf703e-5b00-44b3-ae8b-b655105a073b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.807902] env[62600]: DEBUG oslo_concurrency.lockutils [req-a538f857-2b9f-41c8-adb1-dff0c5ebc3d2 req-77846ae3-e3fb-4445-bfd6-e4922d8de0a3 service nova] Acquired lock "refresh_cache-c6cf703e-5b00-44b3-ae8b-b655105a073b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.808475] env[62600]: DEBUG nova.network.neutron [req-a538f857-2b9f-41c8-adb1-dff0c5ebc3d2 req-77846ae3-e3fb-4445-bfd6-e4922d8de0a3 service nova] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Refreshing network info cache for port 25a47394-9295-4585-8463-7d56a4fba500 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 575.893633] env[62600]: INFO nova.compute.manager [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Rebuilding instance [ 575.948576] env[62600]: DEBUG nova.compute.manager [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 575.949803] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3f347ed-ca87-4384-ba39-bbe50da8c547 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.012130] env[62600]: INFO nova.compute.manager [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] [instance: da07d0e2-98bc-451c-955e-db5eaeee3c82] Took 1.05 seconds to deallocate network for instance. [ 576.240239] env[62600]: DEBUG nova.network.neutron [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.345906] env[62600]: DEBUG nova.network.neutron [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.353514] env[62600]: DEBUG nova.network.neutron [req-a538f857-2b9f-41c8-adb1-dff0c5ebc3d2 req-77846ae3-e3fb-4445-bfd6-e4922d8de0a3 service nova] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.469346] env[62600]: DEBUG nova.network.neutron [req-a538f857-2b9f-41c8-adb1-dff0c5ebc3d2 req-77846ae3-e3fb-4445-bfd6-e4922d8de0a3 service nova] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.471502] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 576.472082] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f24182b6-8acb-4c9c-9d31-2ad86d85be68 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.480456] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Waiting for the task: (returnval){ [ 576.480456] env[62600]: value = "task-1222202" [ 576.480456] env[62600]: _type = "Task" [ 576.480456] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.495426] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Task: {'id': task-1222202, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.564769] env[62600]: DEBUG nova.network.neutron [-] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.767037] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f3ce945-8690-4abd-8541-8dd5529b0811 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.776322] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4175d4e3-af0b-441d-a7c9-d4a3ee6c2d79 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.811197] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee56502b-b046-4349-9fe2-2b7f347c35a8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.825149] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec956643-f534-496a-b3de-998899a1efd2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.843245] env[62600]: DEBUG nova.compute.provider_tree [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.849779] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Releasing lock "refresh_cache-b315463b-f8aa-4b16-a5a2-069458f15358" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.849779] env[62600]: DEBUG nova.compute.manager [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 576.849779] env[62600]: DEBUG nova.compute.manager [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 576.849779] env[62600]: DEBUG nova.network.neutron [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 576.881353] env[62600]: DEBUG nova.network.neutron [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.959265] env[62600]: DEBUG oslo_concurrency.lockutils [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Acquiring lock "60db0e43-692b-4449-9b44-badf3053429d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.959428] env[62600]: DEBUG oslo_concurrency.lockutils [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Lock "60db0e43-692b-4449-9b44-badf3053429d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.973443] env[62600]: DEBUG oslo_concurrency.lockutils [req-a538f857-2b9f-41c8-adb1-dff0c5ebc3d2 req-77846ae3-e3fb-4445-bfd6-e4922d8de0a3 service nova] Releasing lock "refresh_cache-c6cf703e-5b00-44b3-ae8b-b655105a073b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.973753] env[62600]: DEBUG nova.compute.manager [req-a538f857-2b9f-41c8-adb1-dff0c5ebc3d2 req-77846ae3-e3fb-4445-bfd6-e4922d8de0a3 service nova] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Received event network-vif-deleted-25a47394-9295-4585-8463-7d56a4fba500 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 576.991165] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Task: {'id': task-1222202, 'name': PowerOffVM_Task, 'duration_secs': 0.199722} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.991487] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 576.991637] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 576.992406] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7fdc6ed-bd4c-408e-b71e-494f8938805f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.000264] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 577.000264] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-611d6b92-6e32-4394-a10f-68f5cd9976bc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.025210] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 577.025431] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 577.025617] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Deleting the datastore file [datastore1] 98cbcc28-54e0-40de-9feb-6da917806423 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 577.025908] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b81299d3-eb06-430a-97c2-2baf2adc21c1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.034350] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Waiting for the task: (returnval){ [ 577.034350] env[62600]: value = "task-1222204" [ 577.034350] env[62600]: _type = "Task" [ 577.034350] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.045014] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Task: {'id': task-1222204, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.065756] env[62600]: INFO nova.scheduler.client.report [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Deleted allocations for instance da07d0e2-98bc-451c-955e-db5eaeee3c82 [ 577.073573] env[62600]: INFO nova.compute.manager [-] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Took 1.44 seconds to deallocate network for instance. [ 577.076916] env[62600]: DEBUG nova.compute.claims [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 577.076916] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.103912] env[62600]: DEBUG nova.compute.manager [req-91e44a6a-6031-4c53-bdae-172fea5a2b71 req-03441a1b-e629-48d9-b52f-3519066f4e74 service nova] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Received event network-vif-deleted-9c6a925d-932e-4c78-9c8b-aff2cce4382b {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 577.346654] env[62600]: DEBUG nova.scheduler.client.report [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 577.383277] env[62600]: DEBUG nova.network.neutron [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.552399] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Task: {'id': task-1222204, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.087523} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.553092] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 577.553591] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 577.553935] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 577.578743] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec5f1410-e0f5-4588-b59c-e628037b7273 tempest-ServerDiagnosticsTest-767059275 tempest-ServerDiagnosticsTest-767059275-project-member] Lock "da07d0e2-98bc-451c-955e-db5eaeee3c82" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.358s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.852355] env[62600]: DEBUG oslo_concurrency.lockutils [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.170s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.853010] env[62600]: ERROR nova.compute.manager [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a9a15762-111c-4f9c-bd35-b106aec57e3b, please check neutron logs for more information. [ 577.853010] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] Traceback (most recent call last): [ 577.853010] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 577.853010] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] self.driver.spawn(context, instance, image_meta, [ 577.853010] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 577.853010] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 577.853010] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 577.853010] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] vm_ref = self.build_virtual_machine(instance, [ 577.853010] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 577.853010] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] vif_infos = vmwarevif.get_vif_info(self._session, [ 577.853010] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 577.853489] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] for vif in network_info: [ 577.853489] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 577.853489] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] return self._sync_wrapper(fn, *args, **kwargs) [ 577.853489] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 577.853489] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] self.wait() [ 577.853489] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 577.853489] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] self[:] = self._gt.wait() [ 577.853489] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 577.853489] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] return self._exit_event.wait() [ 577.853489] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 577.853489] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] result = hub.switch() [ 577.853489] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 577.853489] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] return self.greenlet.switch() [ 577.853967] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.853967] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] result = function(*args, **kwargs) [ 577.853967] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 577.853967] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] return func(*args, **kwargs) [ 577.853967] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.853967] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] raise e [ 577.853967] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.853967] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] nwinfo = self.network_api.allocate_for_instance( [ 577.853967] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.853967] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] created_port_ids = self._update_ports_for_instance( [ 577.853967] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.853967] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] with excutils.save_and_reraise_exception(): [ 577.853967] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.854560] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] self.force_reraise() [ 577.854560] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.854560] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] raise self.value [ 577.854560] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.854560] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] updated_port = self._update_port( [ 577.854560] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.854560] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] _ensure_no_port_binding_failure(port) [ 577.854560] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.854560] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] raise exception.PortBindingFailed(port_id=port['id']) [ 577.854560] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] nova.exception.PortBindingFailed: Binding failed for port a9a15762-111c-4f9c-bd35-b106aec57e3b, please check neutron logs for more information. [ 577.854560] env[62600]: ERROR nova.compute.manager [instance: 371a3210-1741-43ce-b635-b84314aee77e] [ 577.854982] env[62600]: DEBUG nova.compute.utils [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Binding failed for port a9a15762-111c-4f9c-bd35-b106aec57e3b, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 577.855911] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.975s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.857452] env[62600]: INFO nova.compute.claims [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 577.862210] env[62600]: DEBUG nova.compute.manager [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Build of instance 371a3210-1741-43ce-b635-b84314aee77e was re-scheduled: Binding failed for port a9a15762-111c-4f9c-bd35-b106aec57e3b, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 577.862663] env[62600]: DEBUG nova.compute.manager [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 577.862936] env[62600]: DEBUG oslo_concurrency.lockutils [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Acquiring lock "refresh_cache-371a3210-1741-43ce-b635-b84314aee77e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.863061] env[62600]: DEBUG oslo_concurrency.lockutils [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Acquired lock "refresh_cache-371a3210-1741-43ce-b635-b84314aee77e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.865549] env[62600]: DEBUG nova.network.neutron [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 577.887905] env[62600]: INFO nova.compute.manager [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: b315463b-f8aa-4b16-a5a2-069458f15358] Took 1.04 seconds to deallocate network for instance. [ 578.090959] env[62600]: DEBUG nova.compute.manager [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 578.401681] env[62600]: DEBUG nova.network.neutron [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 578.505704] env[62600]: DEBUG nova.network.neutron [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.609910] env[62600]: DEBUG nova.virt.hardware [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 578.610217] env[62600]: DEBUG nova.virt.hardware [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 578.610401] env[62600]: DEBUG nova.virt.hardware [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 578.610668] env[62600]: DEBUG nova.virt.hardware [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 578.611271] env[62600]: DEBUG nova.virt.hardware [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 578.611271] env[62600]: DEBUG nova.virt.hardware [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 578.611271] env[62600]: DEBUG nova.virt.hardware [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 578.611271] env[62600]: DEBUG nova.virt.hardware [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 578.612664] env[62600]: DEBUG nova.virt.hardware [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 578.612664] env[62600]: DEBUG nova.virt.hardware [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 578.612664] env[62600]: DEBUG nova.virt.hardware [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 578.613481] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c695f1-6629-414b-a37c-f5a186d9b133 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.627415] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e9d9f1-0707-4988-951e-5f431d755832 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.632909] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.647222] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Instance VIF info [] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 578.651338] env[62600]: DEBUG oslo.service.loopingcall [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 578.652038] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 578.652272] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-825385a1-034d-4525-a5d0-afdb9d080c1f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.671210] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 578.671210] env[62600]: value = "task-1222205" [ 578.671210] env[62600]: _type = "Task" [ 578.671210] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.681281] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222205, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.922700] env[62600]: INFO nova.scheduler.client.report [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Deleted allocations for instance b315463b-f8aa-4b16-a5a2-069458f15358 [ 579.007675] env[62600]: DEBUG oslo_concurrency.lockutils [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Releasing lock "refresh_cache-371a3210-1741-43ce-b635-b84314aee77e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.007953] env[62600]: DEBUG nova.compute.manager [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 579.008133] env[62600]: DEBUG nova.compute.manager [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 579.008296] env[62600]: DEBUG nova.network.neutron [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 579.027446] env[62600]: DEBUG nova.network.neutron [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 579.191643] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222205, 'name': CreateVM_Task, 'duration_secs': 0.300464} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 579.192036] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 579.192284] env[62600]: DEBUG oslo_concurrency.lockutils [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.192576] env[62600]: DEBUG oslo_concurrency.lockutils [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.192890] env[62600]: DEBUG oslo_concurrency.lockutils [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 579.193150] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d759395-4089-4bff-a771-f37b35646177 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.199820] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Waiting for the task: (returnval){ [ 579.199820] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d562bc-2cc6-c08f-8394-59cf2abf5eb2" [ 579.199820] env[62600]: _type = "Task" [ 579.199820] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 579.217476] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d562bc-2cc6-c08f-8394-59cf2abf5eb2, 'name': SearchDatastore_Task, 'duration_secs': 0.009253} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 579.220854] env[62600]: DEBUG oslo_concurrency.lockutils [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.222993] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 579.223236] env[62600]: DEBUG oslo_concurrency.lockutils [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.223443] env[62600]: DEBUG oslo_concurrency.lockutils [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.224903] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 579.225411] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e324d21b-ac9b-47d8-8735-2e3cc85267d8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.233753] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 579.233853] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 579.235192] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0744387-3642-4757-9d0b-fc8a0129049f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.245329] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Waiting for the task: (returnval){ [ 579.245329] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c38caa-26e1-5fd9-2a35-8cda022a8366" [ 579.245329] env[62600]: _type = "Task" [ 579.245329] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 579.253486] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c38caa-26e1-5fd9-2a35-8cda022a8366, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.313651] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be7c3fe0-c3b1-4b87-bed4-ed77b46da2bb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.320409] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-561b4b62-e7e7-40e0-b603-ab1a5c491332 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.355360] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0682fdd1-8ec4-4d32-8bbc-fb68902c16b9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.362267] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-768212c4-8ead-43f6-91f3-d51e22ceac46 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.378375] env[62600]: DEBUG nova.compute.provider_tree [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.436686] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ba735759-d4e6-4eaa-85d0-885a66420e3a tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Lock "b315463b-f8aa-4b16-a5a2-069458f15358" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.949s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.531615] env[62600]: DEBUG nova.network.neutron [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.758787] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c38caa-26e1-5fd9-2a35-8cda022a8366, 'name': SearchDatastore_Task, 'duration_secs': 0.008001} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 579.759744] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd62ba24-0bef-4327-8c9a-bac91a45d718 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.765712] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Waiting for the task: (returnval){ [ 579.765712] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]524e5248-8c5f-c982-5a38-493740f02bd2" [ 579.765712] env[62600]: _type = "Task" [ 579.765712] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 579.778799] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]524e5248-8c5f-c982-5a38-493740f02bd2, 'name': SearchDatastore_Task, 'duration_secs': 0.008891} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 579.778799] env[62600]: DEBUG oslo_concurrency.lockutils [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.779074] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 98cbcc28-54e0-40de-9feb-6da917806423/98cbcc28-54e0-40de-9feb-6da917806423.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 579.784260] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-401cd2fc-6755-4c0d-82e8-4f573260e707 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.789166] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Waiting for the task: (returnval){ [ 579.789166] env[62600]: value = "task-1222206" [ 579.789166] env[62600]: _type = "Task" [ 579.789166] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 579.802339] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Task: {'id': task-1222206, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.885440] env[62600]: DEBUG nova.scheduler.client.report [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 579.941853] env[62600]: DEBUG nova.compute.manager [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 580.034043] env[62600]: INFO nova.compute.manager [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] [instance: 371a3210-1741-43ce-b635-b84314aee77e] Took 1.03 seconds to deallocate network for instance. [ 580.303585] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Task: {'id': task-1222206, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458431} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 580.304217] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 98cbcc28-54e0-40de-9feb-6da917806423/98cbcc28-54e0-40de-9feb-6da917806423.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 580.304654] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 580.305028] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-738e3383-78c0-4aaf-9a7c-004615ae1209 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.315158] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Waiting for the task: (returnval){ [ 580.315158] env[62600]: value = "task-1222207" [ 580.315158] env[62600]: _type = "Task" [ 580.315158] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 580.325190] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Task: {'id': task-1222207, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.391767] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.536s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.391767] env[62600]: DEBUG nova.compute.manager [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 580.399441] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.861s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.466800] env[62600]: DEBUG oslo_concurrency.lockutils [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.826547] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Task: {'id': task-1222207, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062996} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 580.826820] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 580.828020] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d90f522a-e073-4acf-b290-fe0c20b4fc98 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.852245] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Reconfiguring VM instance instance-00000001 to attach disk [datastore2] 98cbcc28-54e0-40de-9feb-6da917806423/98cbcc28-54e0-40de-9feb-6da917806423.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 580.852731] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1fb95dd2-3a71-4297-916e-a53aabdf891d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.881672] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Waiting for the task: (returnval){ [ 580.881672] env[62600]: value = "task-1222208" [ 580.881672] env[62600]: _type = "Task" [ 580.881672] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 580.894866] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Task: {'id': task-1222208, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.901459] env[62600]: DEBUG nova.compute.utils [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 580.902803] env[62600]: DEBUG nova.compute.manager [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 580.902953] env[62600]: DEBUG nova.network.neutron [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 581.054523] env[62600]: DEBUG nova.policy [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '76f3dcf16ac34df683548f101b258060', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2141e7e1a0884ed1b512cf12e0d01e55', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 581.099880] env[62600]: INFO nova.scheduler.client.report [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Deleted allocations for instance 371a3210-1741-43ce-b635-b84314aee77e [ 581.381717] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d896165b-c936-4c48-b6a2-92819083f05b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.398253] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Task: {'id': task-1222208, 'name': ReconfigVM_Task, 'duration_secs': 0.283619} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 581.398637] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Reconfigured VM instance instance-00000001 to attach disk [datastore2] 98cbcc28-54e0-40de-9feb-6da917806423/98cbcc28-54e0-40de-9feb-6da917806423.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 581.400091] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817c066e-6963-4628-af63-55dce9a26f25 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.403275] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c0dc3f8e-9d02-4010-a68c-ce8d93b87c2c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.409270] env[62600]: DEBUG nova.compute.manager [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 581.440543] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62a97d9c-0e48-418c-bd85-a4ccb368f1ae {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.444029] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Waiting for the task: (returnval){ [ 581.444029] env[62600]: value = "task-1222209" [ 581.444029] env[62600]: _type = "Task" [ 581.444029] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 581.453028] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-154944dd-59c7-441e-95e8-1e43c61d77a0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.461178] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Task: {'id': task-1222209, 'name': Rename_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.471414] env[62600]: DEBUG nova.compute.provider_tree [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.622763] env[62600]: DEBUG oslo_concurrency.lockutils [None req-852d893d-2bd4-4ff8-be1c-69b5147b7162 tempest-TenantUsagesTestJSON-1365748407 tempest-TenantUsagesTestJSON-1365748407-project-member] Lock "371a3210-1741-43ce-b635-b84314aee77e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.825s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.954880] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Task: {'id': task-1222209, 'name': Rename_Task, 'duration_secs': 0.150247} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 581.955475] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 581.955686] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bca06d61-8fec-49d8-b3c7-a67fa7ca5462 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.962567] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Waiting for the task: (returnval){ [ 581.962567] env[62600]: value = "task-1222210" [ 581.962567] env[62600]: _type = "Task" [ 581.962567] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 581.974521] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Task: {'id': task-1222210, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.975461] env[62600]: DEBUG nova.scheduler.client.report [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 582.128228] env[62600]: DEBUG nova.compute.manager [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 582.419378] env[62600]: DEBUG nova.compute.manager [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 582.453186] env[62600]: DEBUG nova.virt.hardware [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 582.453458] env[62600]: DEBUG nova.virt.hardware [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 582.453609] env[62600]: DEBUG nova.virt.hardware [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 582.453783] env[62600]: DEBUG nova.virt.hardware [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 582.456105] env[62600]: DEBUG nova.virt.hardware [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 582.456388] env[62600]: DEBUG nova.virt.hardware [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 582.456662] env[62600]: DEBUG nova.virt.hardware [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 582.456834] env[62600]: DEBUG nova.virt.hardware [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 582.457016] env[62600]: DEBUG nova.virt.hardware [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 582.457182] env[62600]: DEBUG nova.virt.hardware [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 582.457508] env[62600]: DEBUG nova.virt.hardware [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 582.458461] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bc81ca6-631f-4c3a-9648-336945cbafc3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.477554] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Task: {'id': task-1222210, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.485703] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.086s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.486983] env[62600]: ERROR nova.compute.manager [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port be017d6a-89b0-49f0-8149-7e3eb6b91ef9, please check neutron logs for more information. [ 582.486983] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Traceback (most recent call last): [ 582.486983] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 582.486983] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] self.driver.spawn(context, instance, image_meta, [ 582.486983] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 582.486983] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.486983] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.486983] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] vm_ref = self.build_virtual_machine(instance, [ 582.486983] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.486983] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.486983] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.487366] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] for vif in network_info: [ 582.487366] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 582.487366] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] return self._sync_wrapper(fn, *args, **kwargs) [ 582.487366] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 582.487366] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] self.wait() [ 582.487366] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 582.487366] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] self[:] = self._gt.wait() [ 582.487366] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.487366] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] return self._exit_event.wait() [ 582.487366] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.487366] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] result = hub.switch() [ 582.487366] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.487366] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] return self.greenlet.switch() [ 582.487719] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.487719] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] result = function(*args, **kwargs) [ 582.487719] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.487719] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] return func(*args, **kwargs) [ 582.487719] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.487719] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] raise e [ 582.487719] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.487719] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] nwinfo = self.network_api.allocate_for_instance( [ 582.487719] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 582.487719] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] created_port_ids = self._update_ports_for_instance( [ 582.487719] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 582.487719] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] with excutils.save_and_reraise_exception(): [ 582.487719] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.488077] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] self.force_reraise() [ 582.488077] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.488077] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] raise self.value [ 582.488077] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 582.488077] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] updated_port = self._update_port( [ 582.488077] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.488077] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] _ensure_no_port_binding_failure(port) [ 582.488077] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.488077] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] raise exception.PortBindingFailed(port_id=port['id']) [ 582.488077] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] nova.exception.PortBindingFailed: Binding failed for port be017d6a-89b0-49f0-8149-7e3eb6b91ef9, please check neutron logs for more information. [ 582.488077] env[62600]: ERROR nova.compute.manager [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] [ 582.488390] env[62600]: DEBUG nova.compute.utils [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Binding failed for port be017d6a-89b0-49f0-8149-7e3eb6b91ef9, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 582.489340] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.730s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.494436] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3023c058-40f0-4ffc-a743-0a3f75b2fd08 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.499776] env[62600]: DEBUG nova.compute.manager [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Build of instance 12d03944-78d1-44e4-aaf9-5678f7f3e1ea was re-scheduled: Binding failed for port be017d6a-89b0-49f0-8149-7e3eb6b91ef9, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 582.500190] env[62600]: DEBUG nova.compute.manager [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 582.500840] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Acquiring lock "refresh_cache-12d03944-78d1-44e4-aaf9-5678f7f3e1ea" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.500840] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Acquired lock "refresh_cache-12d03944-78d1-44e4-aaf9-5678f7f3e1ea" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.500840] env[62600]: DEBUG nova.network.neutron [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 582.517484] env[62600]: DEBUG nova.network.neutron [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Successfully created port: 6b748657-b67e-4eb0-86a8-6a8edf4ee8b9 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 582.656662] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.838360] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Acquiring lock "9fd421c6-d068-4a81-a042-2d1f23fb9e1a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.838603] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Lock "9fd421c6-d068-4a81-a042-2d1f23fb9e1a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.982439] env[62600]: DEBUG oslo_vmware.api [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Task: {'id': task-1222210, 'name': PowerOnVM_Task, 'duration_secs': 0.714667} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.982741] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 582.982940] env[62600]: DEBUG nova.compute.manager [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 582.984047] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1adcc203-a945-4714-85b1-febda3138d1b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.055477] env[62600]: DEBUG nova.network.neutron [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 583.233937] env[62600]: DEBUG nova.network.neutron [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.345630] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-490409d6-538f-490a-8ac8-87e1f6a77882 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.354189] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe27030-4a5c-4628-8978-53953c410d79 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.390780] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a4900ef-5aa9-4ef1-9d79-4aae805e8edc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.398968] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f97f65f-e52d-4d0a-b4cb-d2e3735c9859 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.412688] env[62600]: DEBUG nova.compute.provider_tree [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.505154] env[62600]: DEBUG oslo_concurrency.lockutils [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.737208] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Releasing lock "refresh_cache-12d03944-78d1-44e4-aaf9-5678f7f3e1ea" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.737473] env[62600]: DEBUG nova.compute.manager [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 583.737657] env[62600]: DEBUG nova.compute.manager [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 583.738635] env[62600]: DEBUG nova.network.neutron [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 583.870494] env[62600]: DEBUG nova.network.neutron [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 583.918028] env[62600]: DEBUG nova.scheduler.client.report [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 584.373920] env[62600]: DEBUG nova.network.neutron [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.423157] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.934s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.423816] env[62600]: ERROR nova.compute.manager [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a10a7f80-347e-49c9-aae5-76ca002d1a38, please check neutron logs for more information. [ 584.423816] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Traceback (most recent call last): [ 584.423816] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 584.423816] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] self.driver.spawn(context, instance, image_meta, [ 584.423816] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 584.423816] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] self._vmops.spawn(context, instance, image_meta, injected_files, [ 584.423816] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 584.423816] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] vm_ref = self.build_virtual_machine(instance, [ 584.423816] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 584.423816] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] vif_infos = vmwarevif.get_vif_info(self._session, [ 584.423816] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 584.424177] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] for vif in network_info: [ 584.424177] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 584.424177] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] return self._sync_wrapper(fn, *args, **kwargs) [ 584.424177] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 584.424177] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] self.wait() [ 584.424177] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 584.424177] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] self[:] = self._gt.wait() [ 584.424177] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 584.424177] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] return self._exit_event.wait() [ 584.424177] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 584.424177] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] result = hub.switch() [ 584.424177] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 584.424177] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] return self.greenlet.switch() [ 584.424797] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.424797] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] result = function(*args, **kwargs) [ 584.424797] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 584.424797] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] return func(*args, **kwargs) [ 584.424797] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 584.424797] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] raise e [ 584.424797] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.424797] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] nwinfo = self.network_api.allocate_for_instance( [ 584.424797] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 584.424797] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] created_port_ids = self._update_ports_for_instance( [ 584.424797] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 584.424797] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] with excutils.save_and_reraise_exception(): [ 584.424797] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.425180] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] self.force_reraise() [ 584.425180] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.425180] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] raise self.value [ 584.425180] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 584.425180] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] updated_port = self._update_port( [ 584.425180] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.425180] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] _ensure_no_port_binding_failure(port) [ 584.425180] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.425180] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] raise exception.PortBindingFailed(port_id=port['id']) [ 584.425180] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] nova.exception.PortBindingFailed: Binding failed for port a10a7f80-347e-49c9-aae5-76ca002d1a38, please check neutron logs for more information. [ 584.425180] env[62600]: ERROR nova.compute.manager [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] [ 584.425532] env[62600]: DEBUG nova.compute.utils [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Binding failed for port a10a7f80-347e-49c9-aae5-76ca002d1a38, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 584.425742] env[62600]: DEBUG oslo_concurrency.lockutils [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.051s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.428965] env[62600]: DEBUG nova.compute.manager [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Build of instance 5ebc4807-907b-4254-8496-92f1bbb7b761 was re-scheduled: Binding failed for port a10a7f80-347e-49c9-aae5-76ca002d1a38, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 584.431581] env[62600]: DEBUG nova.compute.manager [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 584.432731] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Acquiring lock "refresh_cache-5ebc4807-907b-4254-8496-92f1bbb7b761" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.432731] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Acquired lock "refresh_cache-5ebc4807-907b-4254-8496-92f1bbb7b761" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.432731] env[62600]: DEBUG nova.network.neutron [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 584.878054] env[62600]: INFO nova.compute.manager [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] [instance: 12d03944-78d1-44e4-aaf9-5678f7f3e1ea] Took 1.14 seconds to deallocate network for instance. [ 584.999218] env[62600]: DEBUG nova.network.neutron [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 585.314885] env[62600]: DEBUG nova.network.neutron [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.324552] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d43f5cb-f025-48d3-aece-97dc47d0b8d6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.331802] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e1d3a9c-8f6f-4ea6-9e79-564a8a4201b3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.369161] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5249d27e-b82c-4f27-8c98-3f2e452ec021 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.377337] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-623ef4ed-429e-4cc7-9dc1-5dba34573cfd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.400848] env[62600]: DEBUG nova.compute.provider_tree [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.473127] env[62600]: DEBUG oslo_concurrency.lockutils [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Acquiring lock "98cbcc28-54e0-40de-9feb-6da917806423" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.474155] env[62600]: DEBUG oslo_concurrency.lockutils [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Lock "98cbcc28-54e0-40de-9feb-6da917806423" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.474155] env[62600]: DEBUG oslo_concurrency.lockutils [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Acquiring lock "98cbcc28-54e0-40de-9feb-6da917806423-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.474286] env[62600]: DEBUG oslo_concurrency.lockutils [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Lock "98cbcc28-54e0-40de-9feb-6da917806423-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.474477] env[62600]: DEBUG oslo_concurrency.lockutils [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Lock "98cbcc28-54e0-40de-9feb-6da917806423-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.477661] env[62600]: INFO nova.compute.manager [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Terminating instance [ 585.481666] env[62600]: DEBUG oslo_concurrency.lockutils [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Acquiring lock "refresh_cache-98cbcc28-54e0-40de-9feb-6da917806423" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.481826] env[62600]: DEBUG oslo_concurrency.lockutils [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Acquired lock "refresh_cache-98cbcc28-54e0-40de-9feb-6da917806423" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.481995] env[62600]: DEBUG nova.network.neutron [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 585.820282] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Releasing lock "refresh_cache-5ebc4807-907b-4254-8496-92f1bbb7b761" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.822300] env[62600]: DEBUG nova.compute.manager [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 585.822856] env[62600]: DEBUG nova.compute.manager [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 585.822937] env[62600]: DEBUG nova.network.neutron [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 585.849699] env[62600]: DEBUG nova.network.neutron [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 585.908750] env[62600]: DEBUG nova.scheduler.client.report [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 585.939263] env[62600]: INFO nova.scheduler.client.report [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Deleted allocations for instance 12d03944-78d1-44e4-aaf9-5678f7f3e1ea [ 586.014336] env[62600]: DEBUG nova.network.neutron [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 586.191382] env[62600]: DEBUG nova.network.neutron [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.231856] env[62600]: DEBUG nova.compute.manager [req-c482aa5b-6d6c-48f1-a09e-b9bdb4fcb01d req-d1591259-60df-4daa-bac5-931fd349e0fb service nova] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Received event network-changed-6b748657-b67e-4eb0-86a8-6a8edf4ee8b9 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 586.232059] env[62600]: DEBUG nova.compute.manager [req-c482aa5b-6d6c-48f1-a09e-b9bdb4fcb01d req-d1591259-60df-4daa-bac5-931fd349e0fb service nova] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Refreshing instance network info cache due to event network-changed-6b748657-b67e-4eb0-86a8-6a8edf4ee8b9. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 586.232287] env[62600]: DEBUG oslo_concurrency.lockutils [req-c482aa5b-6d6c-48f1-a09e-b9bdb4fcb01d req-d1591259-60df-4daa-bac5-931fd349e0fb service nova] Acquiring lock "refresh_cache-bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.232432] env[62600]: DEBUG oslo_concurrency.lockutils [req-c482aa5b-6d6c-48f1-a09e-b9bdb4fcb01d req-d1591259-60df-4daa-bac5-931fd349e0fb service nova] Acquired lock "refresh_cache-bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.232611] env[62600]: DEBUG nova.network.neutron [req-c482aa5b-6d6c-48f1-a09e-b9bdb4fcb01d req-d1591259-60df-4daa-bac5-931fd349e0fb service nova] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Refreshing network info cache for port 6b748657-b67e-4eb0-86a8-6a8edf4ee8b9 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 586.321113] env[62600]: ERROR nova.compute.manager [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6b748657-b67e-4eb0-86a8-6a8edf4ee8b9, please check neutron logs for more information. [ 586.321113] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 586.321113] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.321113] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 586.321113] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 586.321113] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 586.321113] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 586.321113] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 586.321113] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.321113] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 586.321113] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.321113] env[62600]: ERROR nova.compute.manager raise self.value [ 586.321113] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 586.321113] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 586.321113] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.321113] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 586.321809] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.321809] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 586.321809] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6b748657-b67e-4eb0-86a8-6a8edf4ee8b9, please check neutron logs for more information. [ 586.321809] env[62600]: ERROR nova.compute.manager [ 586.321809] env[62600]: Traceback (most recent call last): [ 586.321809] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 586.321809] env[62600]: listener.cb(fileno) [ 586.321809] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.321809] env[62600]: result = function(*args, **kwargs) [ 586.321809] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 586.321809] env[62600]: return func(*args, **kwargs) [ 586.321809] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.321809] env[62600]: raise e [ 586.321809] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.321809] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 586.321809] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 586.321809] env[62600]: created_port_ids = self._update_ports_for_instance( [ 586.321809] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 586.321809] env[62600]: with excutils.save_and_reraise_exception(): [ 586.321809] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.321809] env[62600]: self.force_reraise() [ 586.321809] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.321809] env[62600]: raise self.value [ 586.321809] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 586.321809] env[62600]: updated_port = self._update_port( [ 586.321809] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.321809] env[62600]: _ensure_no_port_binding_failure(port) [ 586.321809] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.321809] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 586.323789] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 6b748657-b67e-4eb0-86a8-6a8edf4ee8b9, please check neutron logs for more information. [ 586.323789] env[62600]: Removing descriptor: 16 [ 586.323789] env[62600]: ERROR nova.compute.manager [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6b748657-b67e-4eb0-86a8-6a8edf4ee8b9, please check neutron logs for more information. [ 586.323789] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Traceback (most recent call last): [ 586.323789] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 586.323789] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] yield resources [ 586.323789] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 586.323789] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] self.driver.spawn(context, instance, image_meta, [ 586.323789] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 586.323789] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] self._vmops.spawn(context, instance, image_meta, injected_files, [ 586.323789] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 586.323789] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] vm_ref = self.build_virtual_machine(instance, [ 586.324229] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 586.324229] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] vif_infos = vmwarevif.get_vif_info(self._session, [ 586.324229] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 586.324229] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] for vif in network_info: [ 586.324229] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 586.324229] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] return self._sync_wrapper(fn, *args, **kwargs) [ 586.324229] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 586.324229] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] self.wait() [ 586.324229] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 586.324229] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] self[:] = self._gt.wait() [ 586.324229] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 586.324229] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] return self._exit_event.wait() [ 586.324229] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 586.324674] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] result = hub.switch() [ 586.324674] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 586.324674] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] return self.greenlet.switch() [ 586.324674] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.324674] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] result = function(*args, **kwargs) [ 586.324674] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 586.324674] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] return func(*args, **kwargs) [ 586.324674] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.324674] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] raise e [ 586.324674] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.324674] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] nwinfo = self.network_api.allocate_for_instance( [ 586.324674] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 586.324674] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] created_port_ids = self._update_ports_for_instance( [ 586.325127] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 586.325127] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] with excutils.save_and_reraise_exception(): [ 586.325127] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.325127] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] self.force_reraise() [ 586.325127] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.325127] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] raise self.value [ 586.325127] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 586.325127] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] updated_port = self._update_port( [ 586.325127] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.325127] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] _ensure_no_port_binding_failure(port) [ 586.325127] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.325127] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] raise exception.PortBindingFailed(port_id=port['id']) [ 586.325585] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] nova.exception.PortBindingFailed: Binding failed for port 6b748657-b67e-4eb0-86a8-6a8edf4ee8b9, please check neutron logs for more information. [ 586.325585] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] [ 586.325585] env[62600]: INFO nova.compute.manager [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Terminating instance [ 586.326855] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "refresh_cache-bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.354631] env[62600]: DEBUG nova.network.neutron [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.415100] env[62600]: DEBUG oslo_concurrency.lockutils [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.989s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.415896] env[62600]: ERROR nova.compute.manager [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 987155ef-8896-4776-83a8-4ac68b0e1f18, please check neutron logs for more information. [ 586.415896] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Traceback (most recent call last): [ 586.415896] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 586.415896] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] self.driver.spawn(context, instance, image_meta, [ 586.415896] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 586.415896] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] self._vmops.spawn(context, instance, image_meta, injected_files, [ 586.415896] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 586.415896] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] vm_ref = self.build_virtual_machine(instance, [ 586.415896] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 586.415896] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] vif_infos = vmwarevif.get_vif_info(self._session, [ 586.415896] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 586.416271] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] for vif in network_info: [ 586.416271] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 586.416271] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] return self._sync_wrapper(fn, *args, **kwargs) [ 586.416271] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 586.416271] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] self.wait() [ 586.416271] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 586.416271] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] self[:] = self._gt.wait() [ 586.416271] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 586.416271] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] return self._exit_event.wait() [ 586.416271] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 586.416271] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] result = hub.switch() [ 586.416271] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 586.416271] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] return self.greenlet.switch() [ 586.416664] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.416664] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] result = function(*args, **kwargs) [ 586.416664] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 586.416664] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] return func(*args, **kwargs) [ 586.416664] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.416664] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] raise e [ 586.416664] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.416664] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] nwinfo = self.network_api.allocate_for_instance( [ 586.416664] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 586.416664] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] created_port_ids = self._update_ports_for_instance( [ 586.416664] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 586.416664] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] with excutils.save_and_reraise_exception(): [ 586.416664] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.417150] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] self.force_reraise() [ 586.417150] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.417150] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] raise self.value [ 586.417150] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 586.417150] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] updated_port = self._update_port( [ 586.417150] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.417150] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] _ensure_no_port_binding_failure(port) [ 586.417150] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.417150] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] raise exception.PortBindingFailed(port_id=port['id']) [ 586.417150] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] nova.exception.PortBindingFailed: Binding failed for port 987155ef-8896-4776-83a8-4ac68b0e1f18, please check neutron logs for more information. [ 586.417150] env[62600]: ERROR nova.compute.manager [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] [ 586.417553] env[62600]: DEBUG nova.compute.utils [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Binding failed for port 987155ef-8896-4776-83a8-4ac68b0e1f18, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 586.417974] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.659s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.420502] env[62600]: DEBUG nova.compute.manager [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Build of instance b74bf924-2d0b-4221-9c82-aa5b1ef903ef was re-scheduled: Binding failed for port 987155ef-8896-4776-83a8-4ac68b0e1f18, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 586.420973] env[62600]: DEBUG nova.compute.manager [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 586.421425] env[62600]: DEBUG oslo_concurrency.lockutils [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "refresh_cache-b74bf924-2d0b-4221-9c82-aa5b1ef903ef" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.421574] env[62600]: DEBUG oslo_concurrency.lockutils [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired lock "refresh_cache-b74bf924-2d0b-4221-9c82-aa5b1ef903ef" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.421738] env[62600]: DEBUG nova.network.neutron [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 586.455093] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3d3cf09c-50a7-4517-921e-6aa228a7ed90 tempest-ServerExternalEventsTest-196832025 tempest-ServerExternalEventsTest-196832025-project-member] Lock "12d03944-78d1-44e4-aaf9-5678f7f3e1ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.143s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.693474] env[62600]: DEBUG oslo_concurrency.lockutils [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Releasing lock "refresh_cache-98cbcc28-54e0-40de-9feb-6da917806423" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.693910] env[62600]: DEBUG nova.compute.manager [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 586.694110] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 586.695018] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fdd76af-9b88-4f3f-bba5-89a9e81d6132 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.703411] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 586.703639] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8927212-4559-4bdd-be06-3d5e0fe75c3b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.712887] env[62600]: DEBUG oslo_vmware.api [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 586.712887] env[62600]: value = "task-1222211" [ 586.712887] env[62600]: _type = "Task" [ 586.712887] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.725547] env[62600]: DEBUG oslo_vmware.api [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222211, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.794042] env[62600]: DEBUG nova.network.neutron [req-c482aa5b-6d6c-48f1-a09e-b9bdb4fcb01d req-d1591259-60df-4daa-bac5-931fd349e0fb service nova] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 586.859015] env[62600]: INFO nova.compute.manager [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] [instance: 5ebc4807-907b-4254-8496-92f1bbb7b761] Took 1.04 seconds to deallocate network for instance. [ 586.956856] env[62600]: DEBUG nova.compute.manager [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 586.982866] env[62600]: DEBUG nova.network.neutron [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.059258] env[62600]: DEBUG nova.network.neutron [req-c482aa5b-6d6c-48f1-a09e-b9bdb4fcb01d req-d1591259-60df-4daa-bac5-931fd349e0fb service nova] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.226331] env[62600]: DEBUG oslo_vmware.api [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222211, 'name': PowerOffVM_Task, 'duration_secs': 0.128988} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.226598] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 587.226822] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 587.227157] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-162ff0ed-4313-4523-a06b-be7d77935891 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.249842] env[62600]: DEBUG nova.network.neutron [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.253966] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 587.256024] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 587.256024] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Deleting the datastore file [datastore2] 98cbcc28-54e0-40de-9feb-6da917806423 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 587.256024] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-19c1ae2d-1c1f-4513-8e01-f3c8d6da3bea {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.261891] env[62600]: DEBUG oslo_vmware.api [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for the task: (returnval){ [ 587.261891] env[62600]: value = "task-1222213" [ 587.261891] env[62600]: _type = "Task" [ 587.261891] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.275834] env[62600]: DEBUG oslo_vmware.api [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222213, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.460864] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 98cbcc28-54e0-40de-9feb-6da917806423 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 587.490882] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.564979] env[62600]: DEBUG oslo_concurrency.lockutils [req-c482aa5b-6d6c-48f1-a09e-b9bdb4fcb01d req-d1591259-60df-4daa-bac5-931fd349e0fb service nova] Releasing lock "refresh_cache-bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.565568] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquired lock "refresh_cache-bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.565657] env[62600]: DEBUG nova.network.neutron [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 587.756329] env[62600]: DEBUG oslo_concurrency.lockutils [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Releasing lock "refresh_cache-b74bf924-2d0b-4221-9c82-aa5b1ef903ef" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.756703] env[62600]: DEBUG nova.compute.manager [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 587.756982] env[62600]: DEBUG nova.compute.manager [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.757174] env[62600]: DEBUG nova.network.neutron [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 587.774697] env[62600]: DEBUG oslo_vmware.api [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Task: {'id': task-1222213, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.105479} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.774975] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 587.775127] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 587.775477] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 587.775710] env[62600]: INFO nova.compute.manager [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Took 1.08 seconds to destroy the instance on the hypervisor. [ 587.778022] env[62600]: DEBUG oslo.service.loopingcall [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 587.778022] env[62600]: DEBUG nova.compute.manager [-] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.778022] env[62600]: DEBUG nova.network.neutron [-] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 587.788967] env[62600]: DEBUG nova.network.neutron [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.803802] env[62600]: DEBUG nova.network.neutron [-] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.908797] env[62600]: INFO nova.scheduler.client.report [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Deleted allocations for instance 5ebc4807-907b-4254-8496-92f1bbb7b761 [ 587.927076] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Acquiring lock "e8cbb437-0ace-4af3-a1d9-f96eeeea8504" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.927349] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Lock "e8cbb437-0ace-4af3-a1d9-f96eeeea8504" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.965463] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 5ebc4807-907b-4254-8496-92f1bbb7b761 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 588.100045] env[62600]: DEBUG nova.network.neutron [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 588.236757] env[62600]: DEBUG nova.network.neutron [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.292106] env[62600]: DEBUG nova.network.neutron [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.306781] env[62600]: DEBUG nova.network.neutron [-] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.417247] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54871a0c-b164-44ae-828f-9e01dba1f87a tempest-FloatingIPsAssociationTestJSON-629385621 tempest-FloatingIPsAssociationTestJSON-629385621-project-member] Lock "5ebc4807-907b-4254-8496-92f1bbb7b761" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.906s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.468276] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance b74bf924-2d0b-4221-9c82-aa5b1ef903ef has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 588.468539] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance a664416f-f9f6-4cdd-93b0-8dd63181562e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 588.468586] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 38e57696-33a8-49d8-b061-ae14155ed280 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 588.468689] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance c6cf703e-5b00-44b3-ae8b-b655105a073b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 588.468803] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 588.742826] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Releasing lock "refresh_cache-bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.742826] env[62600]: DEBUG nova.compute.manager [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 588.742826] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 588.742826] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c7656159-3fd1-4ae1-a3e2-d1b54e3a69ea {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.753819] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b93cd795-a3fb-4cd0-80cb-13228144d8b2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.784565] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621 could not be found. [ 588.785565] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 588.785565] env[62600]: INFO nova.compute.manager [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Took 0.04 seconds to destroy the instance on the hypervisor. [ 588.785804] env[62600]: DEBUG oslo.service.loopingcall [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 588.786127] env[62600]: DEBUG nova.compute.manager [-] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 588.786322] env[62600]: DEBUG nova.network.neutron [-] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 588.798036] env[62600]: INFO nova.compute.manager [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: b74bf924-2d0b-4221-9c82-aa5b1ef903ef] Took 1.04 seconds to deallocate network for instance. [ 588.813601] env[62600]: INFO nova.compute.manager [-] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Took 1.04 seconds to deallocate network for instance. [ 588.913103] env[62600]: DEBUG nova.network.neutron [-] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 588.920702] env[62600]: DEBUG nova.compute.manager [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 588.971660] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance ec99a26e-ee0a-44da-80b6-afd1205c7ea0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 589.323021] env[62600]: DEBUG oslo_concurrency.lockutils [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.396166] env[62600]: DEBUG nova.compute.manager [req-e6969f0e-4dfb-4877-a694-604741f5d393 req-3e27d310-1e0c-4e97-9ffc-00797c3b35f0 service nova] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Received event network-vif-deleted-6b748657-b67e-4eb0-86a8-6a8edf4ee8b9 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 589.415779] env[62600]: DEBUG nova.network.neutron [-] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.453302] env[62600]: DEBUG oslo_concurrency.lockutils [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.482273] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 8c0e3a15-0cdd-422a-8a97-95d57621de88 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 589.719865] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Acquiring lock "3d19d349-e450-4307-a763-6269dcb06544" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.720142] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Lock "3d19d349-e450-4307-a763-6269dcb06544" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.844800] env[62600]: INFO nova.scheduler.client.report [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Deleted allocations for instance b74bf924-2d0b-4221-9c82-aa5b1ef903ef [ 589.920683] env[62600]: INFO nova.compute.manager [-] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Took 1.13 seconds to deallocate network for instance. [ 589.924132] env[62600]: DEBUG nova.compute.claims [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 589.924221] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.988057] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance a934e86f-1a51-468e-915f-add473fc2bf6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 590.357874] env[62600]: DEBUG oslo_concurrency.lockutils [None req-649fdb14-fe37-4012-80ca-bd7775b66329 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "b74bf924-2d0b-4221-9c82-aa5b1ef903ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.988s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.492646] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance b3cc82aa-be39-4e89-8972-b5596629961c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 590.869989] env[62600]: DEBUG nova.compute.manager [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 590.997018] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance ba442811-2b0a-44e7-b036-572291aff648 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 591.400141] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.505045] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance b569fb64-27c2-4e20-a2c4-c81058958f98 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 591.734366] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Acquiring lock "932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.734592] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Lock "932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.013085] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance dc77594b-7767-40bf-adbb-c72fa5c80503 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 592.518707] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 8e64b614-31e3-4830-b620-ff8409da8c37 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 592.985063] env[62600]: DEBUG oslo_concurrency.lockutils [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Acquiring lock "20fab902-79cc-4f83-abaf-53a53c2521f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.985867] env[62600]: DEBUG oslo_concurrency.lockutils [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Lock "20fab902-79cc-4f83-abaf-53a53c2521f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.023064] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 6467456c-db39-4fd6-b67e-a5be2b803bd3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 593.533319] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 594.046380] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 222a94b5-7841-490e-9b27-d3f703e6794f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 594.180502] env[62600]: DEBUG oslo_concurrency.lockutils [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquiring lock "6b19c327-9a80-498f-a350-8068ad250377" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.180502] env[62600]: DEBUG oslo_concurrency.lockutils [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "6b19c327-9a80-498f-a350-8068ad250377" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.550241] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 4455c247-a707-4c3a-9323-0110a1290780 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 595.060590] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 09ccf23b-4763-47e1-afad-59e2b06b2d9e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 595.566120] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 60db0e43-692b-4449-9b44-badf3053429d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 595.848689] env[62600]: DEBUG oslo_concurrency.lockutils [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Acquiring lock "15772d8e-98ce-47cf-9b61-e79c7ffc3a8b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.848689] env[62600]: DEBUG oslo_concurrency.lockutils [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Lock "15772d8e-98ce-47cf-9b61-e79c7ffc3a8b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.853690] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "c93cb228-0adb-423c-854f-73354cdb7012" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.853690] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "c93cb228-0adb-423c-854f-73354cdb7012" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.070792] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 9fd421c6-d068-4a81-a042-2d1f23fb9e1a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 596.071020] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 596.071194] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 596.552881] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff48a7d5-3fb4-44f7-8e33-219330cba081 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.567323] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32618561-3856-4902-9fbd-4a60e314a972 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.608495] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbfa7d31-6ef9-4cf6-a350-0dc9d5f097f0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.617777] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-949824ed-13f4-450c-a263-b7ccd773221f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.637310] env[62600]: DEBUG nova.compute.provider_tree [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.147906] env[62600]: DEBUG nova.scheduler.client.report [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.655673] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62600) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 597.655976] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 11.238s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.656255] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.535s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.606383] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3175886a-fc66-432d-8cbe-b7208992d0e1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.615344] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7fb4fbc-079a-4825-a8e7-3f49c8abd45b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.654505] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9484ffa9-2c7e-47a0-bef0-6aeeaae86706 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.663110] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e6bd9c2-e9d1-4e5b-9640-d69ff804b303 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.679669] env[62600]: DEBUG nova.compute.provider_tree [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.183391] env[62600]: DEBUG nova.scheduler.client.report [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.474818] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Acquiring lock "e3416076-1d11-47e4-9f76-9a64d7259166" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.474818] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Lock "e3416076-1d11-47e4-9f76-9a64d7259166" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.689569] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.033s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.690244] env[62600]: ERROR nova.compute.manager [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 60d470c3-2cd9-4861-8309-78911c4b1d06, please check neutron logs for more information. [ 599.690244] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Traceback (most recent call last): [ 599.690244] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 599.690244] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] self.driver.spawn(context, instance, image_meta, [ 599.690244] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 599.690244] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 599.690244] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 599.690244] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] vm_ref = self.build_virtual_machine(instance, [ 599.690244] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 599.690244] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] vif_infos = vmwarevif.get_vif_info(self._session, [ 599.690244] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 599.690658] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] for vif in network_info: [ 599.690658] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 599.690658] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] return self._sync_wrapper(fn, *args, **kwargs) [ 599.690658] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 599.690658] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] self.wait() [ 599.690658] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 599.690658] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] self[:] = self._gt.wait() [ 599.690658] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 599.690658] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] return self._exit_event.wait() [ 599.690658] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 599.690658] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] result = hub.switch() [ 599.690658] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 599.690658] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] return self.greenlet.switch() [ 599.691124] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.691124] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] result = function(*args, **kwargs) [ 599.691124] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 599.691124] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] return func(*args, **kwargs) [ 599.691124] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.691124] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] raise e [ 599.691124] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.691124] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] nwinfo = self.network_api.allocate_for_instance( [ 599.691124] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 599.691124] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] created_port_ids = self._update_ports_for_instance( [ 599.691124] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 599.691124] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] with excutils.save_and_reraise_exception(): [ 599.691124] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.691575] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] self.force_reraise() [ 599.691575] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.691575] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] raise self.value [ 599.691575] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 599.691575] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] updated_port = self._update_port( [ 599.691575] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.691575] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] _ensure_no_port_binding_failure(port) [ 599.691575] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.691575] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] raise exception.PortBindingFailed(port_id=port['id']) [ 599.691575] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] nova.exception.PortBindingFailed: Binding failed for port 60d470c3-2cd9-4861-8309-78911c4b1d06, please check neutron logs for more information. [ 599.691575] env[62600]: ERROR nova.compute.manager [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] [ 599.692941] env[62600]: DEBUG nova.compute.utils [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Binding failed for port 60d470c3-2cd9-4861-8309-78911c4b1d06, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 599.696181] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 24.742s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.696416] env[62600]: DEBUG nova.objects.instance [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62600) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 599.699582] env[62600]: DEBUG nova.compute.manager [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Build of instance a664416f-f9f6-4cdd-93b0-8dd63181562e was re-scheduled: Binding failed for port 60d470c3-2cd9-4861-8309-78911c4b1d06, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 599.699700] env[62600]: DEBUG nova.compute.manager [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 599.700890] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "refresh_cache-a664416f-f9f6-4cdd-93b0-8dd63181562e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.700890] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquired lock "refresh_cache-a664416f-f9f6-4cdd-93b0-8dd63181562e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.700890] env[62600]: DEBUG nova.network.neutron [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 600.224604] env[62600]: DEBUG nova.network.neutron [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.303453] env[62600]: DEBUG nova.network.neutron [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.709023] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c4af7ccc-9527-499f-8536-32fe7185fa7d tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.709023] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.237s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.808027] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Releasing lock "refresh_cache-a664416f-f9f6-4cdd-93b0-8dd63181562e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.808027] env[62600]: DEBUG nova.compute.manager [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 600.808027] env[62600]: DEBUG nova.compute.manager [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 600.808027] env[62600]: DEBUG nova.network.neutron [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 600.831174] env[62600]: DEBUG nova.network.neutron [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.333519] env[62600]: DEBUG nova.network.neutron [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.676332] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e05cfa21-6252-4cad-8fe2-7bf1b17a683d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.684721] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-187cfcd6-d728-4fb1-88bb-acc62504cff7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.718047] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab29fa9f-b85a-4970-b1f8-1b3f51ad3f51 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.724526] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50979659-0d90-4ab9-825b-554b4bb2700d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.743115] env[62600]: DEBUG nova.compute.provider_tree [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.836471] env[62600]: INFO nova.compute.manager [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: a664416f-f9f6-4cdd-93b0-8dd63181562e] Took 1.03 seconds to deallocate network for instance. [ 602.246820] env[62600]: DEBUG nova.scheduler.client.report [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 602.751664] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.043s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.755280] env[62600]: ERROR nova.compute.manager [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9c6a925d-932e-4c78-9c8b-aff2cce4382b, please check neutron logs for more information. [ 602.755280] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Traceback (most recent call last): [ 602.755280] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 602.755280] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] self.driver.spawn(context, instance, image_meta, [ 602.755280] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 602.755280] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.755280] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.755280] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] vm_ref = self.build_virtual_machine(instance, [ 602.755280] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.755280] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.755280] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.756938] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] for vif in network_info: [ 602.756938] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.756938] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] return self._sync_wrapper(fn, *args, **kwargs) [ 602.756938] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.756938] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] self.wait() [ 602.756938] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.756938] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] self[:] = self._gt.wait() [ 602.756938] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.756938] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] return self._exit_event.wait() [ 602.756938] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.756938] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] result = hub.switch() [ 602.756938] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.756938] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] return self.greenlet.switch() [ 602.757730] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.757730] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] result = function(*args, **kwargs) [ 602.757730] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.757730] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] return func(*args, **kwargs) [ 602.757730] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.757730] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] raise e [ 602.757730] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.757730] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] nwinfo = self.network_api.allocate_for_instance( [ 602.757730] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.757730] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] created_port_ids = self._update_ports_for_instance( [ 602.757730] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.757730] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] with excutils.save_and_reraise_exception(): [ 602.757730] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.758168] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] self.force_reraise() [ 602.758168] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.758168] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] raise self.value [ 602.758168] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.758168] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] updated_port = self._update_port( [ 602.758168] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.758168] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] _ensure_no_port_binding_failure(port) [ 602.758168] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.758168] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] raise exception.PortBindingFailed(port_id=port['id']) [ 602.758168] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] nova.exception.PortBindingFailed: Binding failed for port 9c6a925d-932e-4c78-9c8b-aff2cce4382b, please check neutron logs for more information. [ 602.758168] env[62600]: ERROR nova.compute.manager [instance: 38e57696-33a8-49d8-b061-ae14155ed280] [ 602.758955] env[62600]: DEBUG nova.compute.utils [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Binding failed for port 9c6a925d-932e-4c78-9c8b-aff2cce4382b, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 602.761047] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.684s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.763912] env[62600]: DEBUG nova.compute.manager [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Build of instance 38e57696-33a8-49d8-b061-ae14155ed280 was re-scheduled: Binding failed for port 9c6a925d-932e-4c78-9c8b-aff2cce4382b, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 602.764951] env[62600]: DEBUG nova.compute.manager [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 602.765783] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Acquiring lock "refresh_cache-38e57696-33a8-49d8-b061-ae14155ed280" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.765783] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Acquired lock "refresh_cache-38e57696-33a8-49d8-b061-ae14155ed280" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.765783] env[62600]: DEBUG nova.network.neutron [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 602.873716] env[62600]: INFO nova.scheduler.client.report [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Deleted allocations for instance a664416f-f9f6-4cdd-93b0-8dd63181562e [ 603.301876] env[62600]: DEBUG nova.network.neutron [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.384809] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e3a583bf-b5c3-41c7-8be5-c64c664bb2d1 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "a664416f-f9f6-4cdd-93b0-8dd63181562e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.781s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.406358] env[62600]: DEBUG nova.network.neutron [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.448281] env[62600]: DEBUG oslo_concurrency.lockutils [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Acquiring lock "877f5b67-68f8-48b0-8e12-622d9b50a390" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.448510] env[62600]: DEBUG oslo_concurrency.lockutils [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Lock "877f5b67-68f8-48b0-8e12-622d9b50a390" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.704163] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b2d41f-9641-431b-a907-faeb2a87b580 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.711148] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8ed3296-9878-4d93-ac0b-644ae3d6b119 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.747544] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2e0e133-2c4d-4eb2-aa10-fc8f87741f3d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.755153] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56ab10c4-699c-4db9-bf20-d797ffa92c50 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.772189] env[62600]: DEBUG nova.compute.provider_tree [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.890453] env[62600]: DEBUG nova.compute.manager [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 603.915029] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Releasing lock "refresh_cache-38e57696-33a8-49d8-b061-ae14155ed280" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.915029] env[62600]: DEBUG nova.compute.manager [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 603.915029] env[62600]: DEBUG nova.compute.manager [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 603.915029] env[62600]: DEBUG nova.network.neutron [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 603.936665] env[62600]: DEBUG nova.network.neutron [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.275327] env[62600]: DEBUG nova.scheduler.client.report [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.417404] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.444285] env[62600]: DEBUG nova.network.neutron [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.784309] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.024s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.784933] env[62600]: ERROR nova.compute.manager [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 25a47394-9295-4585-8463-7d56a4fba500, please check neutron logs for more information. [ 604.784933] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Traceback (most recent call last): [ 604.784933] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.784933] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] self.driver.spawn(context, instance, image_meta, [ 604.784933] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 604.784933] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.784933] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.784933] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] vm_ref = self.build_virtual_machine(instance, [ 604.784933] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.784933] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.784933] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.785456] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] for vif in network_info: [ 604.785456] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 604.785456] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] return self._sync_wrapper(fn, *args, **kwargs) [ 604.785456] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 604.785456] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] self.wait() [ 604.785456] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 604.785456] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] self[:] = self._gt.wait() [ 604.785456] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.785456] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] return self._exit_event.wait() [ 604.785456] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 604.785456] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] result = hub.switch() [ 604.785456] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 604.785456] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] return self.greenlet.switch() [ 604.785973] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.785973] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] result = function(*args, **kwargs) [ 604.785973] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 604.785973] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] return func(*args, **kwargs) [ 604.785973] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.785973] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] raise e [ 604.785973] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.785973] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] nwinfo = self.network_api.allocate_for_instance( [ 604.785973] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 604.785973] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] created_port_ids = self._update_ports_for_instance( [ 604.785973] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 604.785973] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] with excutils.save_and_reraise_exception(): [ 604.785973] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.786526] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] self.force_reraise() [ 604.786526] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.786526] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] raise self.value [ 604.786526] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 604.786526] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] updated_port = self._update_port( [ 604.786526] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.786526] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] _ensure_no_port_binding_failure(port) [ 604.786526] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.786526] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] raise exception.PortBindingFailed(port_id=port['id']) [ 604.786526] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] nova.exception.PortBindingFailed: Binding failed for port 25a47394-9295-4585-8463-7d56a4fba500, please check neutron logs for more information. [ 604.786526] env[62600]: ERROR nova.compute.manager [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] [ 604.786933] env[62600]: DEBUG nova.compute.utils [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Binding failed for port 25a47394-9295-4585-8463-7d56a4fba500, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 604.787186] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.155s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.788678] env[62600]: INFO nova.compute.claims [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 604.792024] env[62600]: DEBUG nova.compute.manager [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Build of instance c6cf703e-5b00-44b3-ae8b-b655105a073b was re-scheduled: Binding failed for port 25a47394-9295-4585-8463-7d56a4fba500, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 604.792024] env[62600]: DEBUG nova.compute.manager [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 604.792024] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Acquiring lock "refresh_cache-c6cf703e-5b00-44b3-ae8b-b655105a073b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.792253] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Acquired lock "refresh_cache-c6cf703e-5b00-44b3-ae8b-b655105a073b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.792310] env[62600]: DEBUG nova.network.neutron [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 604.948280] env[62600]: INFO nova.compute.manager [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] [instance: 38e57696-33a8-49d8-b061-ae14155ed280] Took 1.03 seconds to deallocate network for instance. [ 605.099306] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.099562] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.318369] env[62600]: DEBUG nova.network.neutron [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.438064] env[62600]: DEBUG nova.network.neutron [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.944769] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Releasing lock "refresh_cache-c6cf703e-5b00-44b3-ae8b-b655105a073b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.945093] env[62600]: DEBUG nova.compute.manager [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 605.946082] env[62600]: DEBUG nova.compute.manager [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 605.946082] env[62600]: DEBUG nova.network.neutron [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 605.965160] env[62600]: DEBUG nova.network.neutron [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.982124] env[62600]: INFO nova.scheduler.client.report [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Deleted allocations for instance 38e57696-33a8-49d8-b061-ae14155ed280 [ 606.212024] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquiring lock "f989a4e4-c3c6-4cb5-9464-cbfb9d66c202" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.212276] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Lock "f989a4e4-c3c6-4cb5-9464-cbfb9d66c202" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.244436] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquiring lock "9c2c8dd4-85c7-489a-95c0-669644237b8b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.244664] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Lock "9c2c8dd4-85c7-489a-95c0-669644237b8b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.276608] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquiring lock "81814471-a617-4364-a140-105252ca9c04" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.276849] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Lock "81814471-a617-4364-a140-105252ca9c04" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.289140] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54800fdb-9f97-4901-92c3-fb330ca582b7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.301150] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f042ba-a3aa-45ce-9ffb-918b706a578a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.341400] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5513740c-f9d1-4712-8dbe-46a4fdffd871 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.350029] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf3d97f3-5d56-43dd-8e40-c728c3b763c8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.364046] env[62600]: DEBUG nova.compute.provider_tree [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.468423] env[62600]: DEBUG nova.network.neutron [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.490829] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9247844-e03f-491e-b55b-e539f079e264 tempest-ServersAdminNegativeTestJSON-134810259 tempest-ServersAdminNegativeTestJSON-134810259-project-member] Lock "38e57696-33a8-49d8-b061-ae14155ed280" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.305s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.866637] env[62600]: DEBUG nova.scheduler.client.report [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.974150] env[62600]: INFO nova.compute.manager [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: c6cf703e-5b00-44b3-ae8b-b655105a073b] Took 1.03 seconds to deallocate network for instance. [ 606.995375] env[62600]: DEBUG nova.compute.manager [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 607.375986] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.589s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.376815] env[62600]: DEBUG nova.compute.manager [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 607.381349] env[62600]: DEBUG oslo_concurrency.lockutils [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.914s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.384180] env[62600]: INFO nova.compute.claims [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 607.530699] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.889221] env[62600]: DEBUG nova.compute.utils [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 607.892698] env[62600]: DEBUG nova.compute.manager [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 607.893651] env[62600]: DEBUG nova.network.neutron [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 607.957235] env[62600]: DEBUG nova.policy [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4484efbbf22c44e9ac4b2e56310bceec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '52e0063fd8ef44a68a4cfee5f759920e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 608.026877] env[62600]: INFO nova.scheduler.client.report [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Deleted allocations for instance c6cf703e-5b00-44b3-ae8b-b655105a073b [ 608.398996] env[62600]: DEBUG nova.compute.manager [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 608.413948] env[62600]: DEBUG nova.network.neutron [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Successfully created port: 660d01e6-2e9a-4002-8ee8-c9b96f07cec5 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 608.540931] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b1854fe-8b7b-4435-8454-53f27eb243c5 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Lock "c6cf703e-5b00-44b3-ae8b-b655105a073b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.019s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.945015] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04adbda7-7186-4a7d-88d0-c8fb6ada801c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.955424] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d762d1-43fc-4d65-8612-ca11393c7128 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.014642] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c1daaf3-a193-438e-97c0-3cc0c5cb011a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.023970] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c695c807-02d0-4c52-8c42-8db226cffc49 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.039656] env[62600]: DEBUG nova.compute.provider_tree [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 609.050826] env[62600]: DEBUG nova.compute.manager [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 609.411039] env[62600]: DEBUG nova.compute.manager [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 609.444906] env[62600]: DEBUG nova.virt.hardware [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 609.445161] env[62600]: DEBUG nova.virt.hardware [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 609.445340] env[62600]: DEBUG nova.virt.hardware [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 609.445541] env[62600]: DEBUG nova.virt.hardware [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 609.445688] env[62600]: DEBUG nova.virt.hardware [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 609.445836] env[62600]: DEBUG nova.virt.hardware [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 609.446252] env[62600]: DEBUG nova.virt.hardware [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 609.446500] env[62600]: DEBUG nova.virt.hardware [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 609.446694] env[62600]: DEBUG nova.virt.hardware [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 609.446862] env[62600]: DEBUG nova.virt.hardware [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 609.447047] env[62600]: DEBUG nova.virt.hardware [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 609.447969] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f43833e1-6d13-46f6-95bb-54b97a0ad052 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.456883] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ed824ae-d362-4264-b804-206d2174ba51 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.543278] env[62600]: DEBUG nova.scheduler.client.report [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.582891] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.806135] env[62600]: DEBUG nova.compute.manager [req-9099d31a-2374-4afc-85ba-b2149b27ccf6 req-4860b4b3-9d59-49f2-9ad8-defeee921efd service nova] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Received event network-changed-660d01e6-2e9a-4002-8ee8-c9b96f07cec5 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 609.806341] env[62600]: DEBUG nova.compute.manager [req-9099d31a-2374-4afc-85ba-b2149b27ccf6 req-4860b4b3-9d59-49f2-9ad8-defeee921efd service nova] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Refreshing instance network info cache due to event network-changed-660d01e6-2e9a-4002-8ee8-c9b96f07cec5. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 609.806558] env[62600]: DEBUG oslo_concurrency.lockutils [req-9099d31a-2374-4afc-85ba-b2149b27ccf6 req-4860b4b3-9d59-49f2-9ad8-defeee921efd service nova] Acquiring lock "refresh_cache-ec99a26e-ee0a-44da-80b6-afd1205c7ea0" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.806701] env[62600]: DEBUG oslo_concurrency.lockutils [req-9099d31a-2374-4afc-85ba-b2149b27ccf6 req-4860b4b3-9d59-49f2-9ad8-defeee921efd service nova] Acquired lock "refresh_cache-ec99a26e-ee0a-44da-80b6-afd1205c7ea0" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.806861] env[62600]: DEBUG nova.network.neutron [req-9099d31a-2374-4afc-85ba-b2149b27ccf6 req-4860b4b3-9d59-49f2-9ad8-defeee921efd service nova] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Refreshing network info cache for port 660d01e6-2e9a-4002-8ee8-c9b96f07cec5 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 610.053479] env[62600]: DEBUG oslo_concurrency.lockutils [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.670s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.053479] env[62600]: DEBUG nova.compute.manager [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 610.057147] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.397s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.057147] env[62600]: INFO nova.compute.claims [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 610.060245] env[62600]: ERROR nova.compute.manager [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 660d01e6-2e9a-4002-8ee8-c9b96f07cec5, please check neutron logs for more information. [ 610.060245] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 610.060245] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.060245] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 610.060245] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 610.060245] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 610.060245] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 610.060245] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 610.060245] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.060245] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 610.060245] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.060245] env[62600]: ERROR nova.compute.manager raise self.value [ 610.060245] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 610.060245] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 610.060245] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.060245] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 610.060846] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.060846] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 610.060846] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 660d01e6-2e9a-4002-8ee8-c9b96f07cec5, please check neutron logs for more information. [ 610.060846] env[62600]: ERROR nova.compute.manager [ 610.060846] env[62600]: Traceback (most recent call last): [ 610.060846] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 610.060846] env[62600]: listener.cb(fileno) [ 610.060846] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.060846] env[62600]: result = function(*args, **kwargs) [ 610.060846] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 610.060846] env[62600]: return func(*args, **kwargs) [ 610.060846] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.060846] env[62600]: raise e [ 610.060846] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.060846] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 610.060846] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 610.060846] env[62600]: created_port_ids = self._update_ports_for_instance( [ 610.060846] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 610.060846] env[62600]: with excutils.save_and_reraise_exception(): [ 610.060846] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.060846] env[62600]: self.force_reraise() [ 610.060846] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.060846] env[62600]: raise self.value [ 610.060846] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 610.060846] env[62600]: updated_port = self._update_port( [ 610.060846] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.060846] env[62600]: _ensure_no_port_binding_failure(port) [ 610.060846] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.060846] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 610.061719] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 660d01e6-2e9a-4002-8ee8-c9b96f07cec5, please check neutron logs for more information. [ 610.061719] env[62600]: Removing descriptor: 16 [ 610.061719] env[62600]: ERROR nova.compute.manager [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 660d01e6-2e9a-4002-8ee8-c9b96f07cec5, please check neutron logs for more information. [ 610.061719] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Traceback (most recent call last): [ 610.061719] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 610.061719] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] yield resources [ 610.061719] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.061719] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] self.driver.spawn(context, instance, image_meta, [ 610.061719] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 610.061719] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.061719] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.061719] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] vm_ref = self.build_virtual_machine(instance, [ 610.062199] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.062199] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.062199] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.062199] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] for vif in network_info: [ 610.062199] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 610.062199] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] return self._sync_wrapper(fn, *args, **kwargs) [ 610.062199] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 610.062199] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] self.wait() [ 610.062199] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 610.062199] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] self[:] = self._gt.wait() [ 610.062199] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.062199] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] return self._exit_event.wait() [ 610.062199] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 610.062616] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] result = hub.switch() [ 610.062616] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 610.062616] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] return self.greenlet.switch() [ 610.062616] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.062616] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] result = function(*args, **kwargs) [ 610.062616] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 610.062616] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] return func(*args, **kwargs) [ 610.062616] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.062616] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] raise e [ 610.062616] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.062616] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] nwinfo = self.network_api.allocate_for_instance( [ 610.062616] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 610.062616] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] created_port_ids = self._update_ports_for_instance( [ 610.063061] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 610.063061] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] with excutils.save_and_reraise_exception(): [ 610.063061] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.063061] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] self.force_reraise() [ 610.063061] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.063061] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] raise self.value [ 610.063061] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 610.063061] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] updated_port = self._update_port( [ 610.063061] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.063061] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] _ensure_no_port_binding_failure(port) [ 610.063061] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.063061] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] raise exception.PortBindingFailed(port_id=port['id']) [ 610.063554] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] nova.exception.PortBindingFailed: Binding failed for port 660d01e6-2e9a-4002-8ee8-c9b96f07cec5, please check neutron logs for more information. [ 610.063554] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] [ 610.063554] env[62600]: INFO nova.compute.manager [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Terminating instance [ 610.064082] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Acquiring lock "refresh_cache-ec99a26e-ee0a-44da-80b6-afd1205c7ea0" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.331233] env[62600]: DEBUG nova.network.neutron [req-9099d31a-2374-4afc-85ba-b2149b27ccf6 req-4860b4b3-9d59-49f2-9ad8-defeee921efd service nova] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.459849] env[62600]: DEBUG nova.network.neutron [req-9099d31a-2374-4afc-85ba-b2149b27ccf6 req-4860b4b3-9d59-49f2-9ad8-defeee921efd service nova] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.564285] env[62600]: DEBUG nova.compute.utils [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 610.565412] env[62600]: DEBUG nova.compute.manager [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Not allocating networking since 'none' was specified. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 610.962669] env[62600]: DEBUG oslo_concurrency.lockutils [req-9099d31a-2374-4afc-85ba-b2149b27ccf6 req-4860b4b3-9d59-49f2-9ad8-defeee921efd service nova] Releasing lock "refresh_cache-ec99a26e-ee0a-44da-80b6-afd1205c7ea0" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.965137] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Acquired lock "refresh_cache-ec99a26e-ee0a-44da-80b6-afd1205c7ea0" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.965137] env[62600]: DEBUG nova.network.neutron [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 611.069195] env[62600]: DEBUG nova.compute.manager [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 611.491056] env[62600]: DEBUG nova.network.neutron [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.505701] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5344e0ca-6395-4b2f-ae3b-ceb74d6a56aa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.514606] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8b5d48-2729-404f-bbe4-7b14f9ee3acb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.563303] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ec10d5-b12d-43db-8548-1553a3c78de6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.573783] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdcfb40f-8058-4f83-94d4-100b082cc65d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.594441] env[62600]: DEBUG nova.compute.provider_tree [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.618203] env[62600]: DEBUG nova.network.neutron [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.946832] env[62600]: DEBUG nova.compute.manager [req-fc845f9c-b25a-4740-bfa2-570c73daf1e2 req-99956c88-a9c9-4acf-b5d5-dcb127678983 service nova] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Received event network-vif-deleted-660d01e6-2e9a-4002-8ee8-c9b96f07cec5 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 612.098197] env[62600]: DEBUG nova.compute.manager [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 612.100167] env[62600]: DEBUG nova.scheduler.client.report [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 612.126185] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Releasing lock "refresh_cache-ec99a26e-ee0a-44da-80b6-afd1205c7ea0" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.126185] env[62600]: DEBUG nova.compute.manager [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 612.126185] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 612.126185] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d850f551-d859-4c6d-bab1-7533c885be05 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.142910] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-187058c3-c0ec-42d1-a873-854d8aa7176b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.172537] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ec99a26e-ee0a-44da-80b6-afd1205c7ea0 could not be found. [ 612.172729] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 612.172910] env[62600]: INFO nova.compute.manager [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Took 0.05 seconds to destroy the instance on the hypervisor. [ 612.173171] env[62600]: DEBUG oslo.service.loopingcall [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 612.174041] env[62600]: DEBUG nova.compute.manager [-] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.174041] env[62600]: DEBUG nova.network.neutron [-] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 612.180119] env[62600]: DEBUG nova.virt.hardware [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 612.180119] env[62600]: DEBUG nova.virt.hardware [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 612.180119] env[62600]: DEBUG nova.virt.hardware [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 612.180403] env[62600]: DEBUG nova.virt.hardware [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 612.180403] env[62600]: DEBUG nova.virt.hardware [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 612.180403] env[62600]: DEBUG nova.virt.hardware [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 612.180403] env[62600]: DEBUG nova.virt.hardware [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 612.180403] env[62600]: DEBUG nova.virt.hardware [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 612.180690] env[62600]: DEBUG nova.virt.hardware [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 612.180690] env[62600]: DEBUG nova.virt.hardware [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 612.180853] env[62600]: DEBUG nova.virt.hardware [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 612.181956] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-412ef778-fddf-451f-978b-a2856d618653 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.189364] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-272e7d9f-32b0-41e5-9275-aa69b4a8c451 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.194240] env[62600]: DEBUG nova.network.neutron [-] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 612.209970] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Instance VIF info [] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 612.217125] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Creating folder: Project (779e2ee67b5e4cc4a365377c0118a222). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 612.217125] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-315f970f-518b-4889-aa1e-a56127dbc8a6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.229798] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Created folder: Project (779e2ee67b5e4cc4a365377c0118a222) in parent group-v264198. [ 612.230018] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Creating folder: Instances. Parent ref: group-v264211. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 612.230252] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6b8a9968-92b7-435d-9fce-da3a286f454e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.239940] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Created folder: Instances in parent group-v264211. [ 612.240206] env[62600]: DEBUG oslo.service.loopingcall [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 612.240385] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 612.240775] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8f8643ce-fdea-4815-9411-9ef1c3f1ae06 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.257271] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 612.257271] env[62600]: value = "task-1222230" [ 612.257271] env[62600]: _type = "Task" [ 612.257271] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.264522] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222230, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 612.608091] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.554s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.608091] env[62600]: DEBUG nova.compute.manager [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 612.614816] env[62600]: DEBUG oslo_concurrency.lockutils [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 29.107s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.614816] env[62600]: DEBUG nova.objects.instance [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62600) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 612.700874] env[62600]: DEBUG nova.network.neutron [-] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.768657] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Acquiring lock "68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.768958] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Lock "68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.773600] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222230, 'name': CreateVM_Task, 'duration_secs': 0.25892} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.774203] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 612.774542] env[62600]: DEBUG oslo_concurrency.lockutils [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.774739] env[62600]: DEBUG oslo_concurrency.lockutils [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.775021] env[62600]: DEBUG oslo_concurrency.lockutils [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 612.775366] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-922b56ae-532f-442f-9c05-431bca94226e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.781794] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Waiting for the task: (returnval){ [ 612.781794] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52a5a62a-9fc6-a968-5e9b-027bb444870e" [ 612.781794] env[62600]: _type = "Task" [ 612.781794] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 612.794020] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52a5a62a-9fc6-a968-5e9b-027bb444870e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.114287] env[62600]: DEBUG nova.compute.utils [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 613.116025] env[62600]: DEBUG nova.compute.manager [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 613.116610] env[62600]: DEBUG nova.network.neutron [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 613.166055] env[62600]: DEBUG nova.policy [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ecef878e2d494b47b76eeeee20abe886', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'abc5d9b5781247a69855b11a1c71aed2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 613.206921] env[62600]: INFO nova.compute.manager [-] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Took 1.03 seconds to deallocate network for instance. [ 613.212198] env[62600]: DEBUG nova.compute.claims [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 613.212563] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.298768] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52a5a62a-9fc6-a968-5e9b-027bb444870e, 'name': SearchDatastore_Task, 'duration_secs': 0.012873} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.299202] env[62600]: DEBUG oslo_concurrency.lockutils [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.300529] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 613.300529] env[62600]: DEBUG oslo_concurrency.lockutils [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.300529] env[62600]: DEBUG oslo_concurrency.lockutils [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.300529] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 613.300529] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e77f7b78-54f5-4ce7-a0ca-f509d6c450a1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.314720] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 613.314720] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 613.315623] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99501561-bf33-4492-adb3-4584df26d0c6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.328205] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Waiting for the task: (returnval){ [ 613.328205] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520ac91a-ed78-4249-fb46-5a16c3f3a798" [ 613.328205] env[62600]: _type = "Task" [ 613.328205] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.340736] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520ac91a-ed78-4249-fb46-5a16c3f3a798, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 613.573361] env[62600]: DEBUG nova.network.neutron [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Successfully created port: 81ec8210-343f-40b7-b92e-ad5a81d2b07e {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 613.621264] env[62600]: DEBUG nova.compute.manager [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 613.631581] env[62600]: DEBUG oslo_concurrency.lockutils [None req-072c90ec-1a0b-4540-b946-bbf0b4aed855 tempest-ServersAdmin275Test-227002365 tempest-ServersAdmin275Test-227002365-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.633252] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.142s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.635993] env[62600]: INFO nova.compute.claims [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 613.846392] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520ac91a-ed78-4249-fb46-5a16c3f3a798, 'name': SearchDatastore_Task, 'duration_secs': 0.016915} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 613.847639] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6efabdcb-1f45-45e8-b8b1-41c3eaa7049a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.853152] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Waiting for the task: (returnval){ [ 613.853152] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52054dc6-3287-4431-6be4-3276c45c2047" [ 613.853152] env[62600]: _type = "Task" [ 613.853152] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 613.860937] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52054dc6-3287-4431-6be4-3276c45c2047, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.363961] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52054dc6-3287-4431-6be4-3276c45c2047, 'name': SearchDatastore_Task, 'duration_secs': 0.021999} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.364328] env[62600]: DEBUG oslo_concurrency.lockutils [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.364526] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 8c0e3a15-0cdd-422a-8a97-95d57621de88/8c0e3a15-0cdd-422a-8a97-95d57621de88.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 614.364777] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bfb8fbef-e513-42da-9b1f-566b4f6d45c7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.374595] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Waiting for the task: (returnval){ [ 614.374595] env[62600]: value = "task-1222232" [ 614.374595] env[62600]: _type = "Task" [ 614.374595] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.384821] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': task-1222232, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 614.429020] env[62600]: DEBUG oslo_concurrency.lockutils [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Acquiring lock "606becc8-ff43-469c-aca6-c35d67614960" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.429020] env[62600]: DEBUG oslo_concurrency.lockutils [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Lock "606becc8-ff43-469c-aca6-c35d67614960" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.639321] env[62600]: DEBUG nova.compute.manager [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 614.672048] env[62600]: DEBUG nova.virt.hardware [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 614.673670] env[62600]: DEBUG nova.virt.hardware [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 614.673670] env[62600]: DEBUG nova.virt.hardware [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 614.673670] env[62600]: DEBUG nova.virt.hardware [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 614.673670] env[62600]: DEBUG nova.virt.hardware [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 614.673670] env[62600]: DEBUG nova.virt.hardware [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 614.673930] env[62600]: DEBUG nova.virt.hardware [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 614.673930] env[62600]: DEBUG nova.virt.hardware [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 614.673930] env[62600]: DEBUG nova.virt.hardware [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 614.673930] env[62600]: DEBUG nova.virt.hardware [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 614.673930] env[62600]: DEBUG nova.virt.hardware [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 614.674948] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ac22f69-3035-4a01-9a25-b4eeb242759d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.691444] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-182d60fc-e2dc-46f2-9af1-594c2fc4c874 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.754266] env[62600]: ERROR nova.compute.manager [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 81ec8210-343f-40b7-b92e-ad5a81d2b07e, please check neutron logs for more information. [ 614.754266] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 614.754266] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.754266] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 614.754266] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 614.754266] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 614.754266] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 614.754266] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 614.754266] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.754266] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 614.754266] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.754266] env[62600]: ERROR nova.compute.manager raise self.value [ 614.754266] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 614.754266] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 614.754266] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.754266] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 614.754874] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.754874] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 614.754874] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 81ec8210-343f-40b7-b92e-ad5a81d2b07e, please check neutron logs for more information. [ 614.754874] env[62600]: ERROR nova.compute.manager [ 614.754874] env[62600]: Traceback (most recent call last): [ 614.754874] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 614.754874] env[62600]: listener.cb(fileno) [ 614.754874] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.754874] env[62600]: result = function(*args, **kwargs) [ 614.754874] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.754874] env[62600]: return func(*args, **kwargs) [ 614.754874] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.754874] env[62600]: raise e [ 614.754874] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.754874] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 614.754874] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 614.754874] env[62600]: created_port_ids = self._update_ports_for_instance( [ 614.754874] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 614.754874] env[62600]: with excutils.save_and_reraise_exception(): [ 614.754874] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.754874] env[62600]: self.force_reraise() [ 614.754874] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.754874] env[62600]: raise self.value [ 614.754874] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 614.754874] env[62600]: updated_port = self._update_port( [ 614.754874] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.754874] env[62600]: _ensure_no_port_binding_failure(port) [ 614.754874] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.754874] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 614.755896] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 81ec8210-343f-40b7-b92e-ad5a81d2b07e, please check neutron logs for more information. [ 614.755896] env[62600]: Removing descriptor: 16 [ 614.755896] env[62600]: ERROR nova.compute.manager [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 81ec8210-343f-40b7-b92e-ad5a81d2b07e, please check neutron logs for more information. [ 614.755896] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Traceback (most recent call last): [ 614.755896] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 614.755896] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] yield resources [ 614.755896] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 614.755896] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] self.driver.spawn(context, instance, image_meta, [ 614.755896] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 614.755896] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 614.755896] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 614.755896] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] vm_ref = self.build_virtual_machine(instance, [ 614.756329] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 614.756329] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] vif_infos = vmwarevif.get_vif_info(self._session, [ 614.756329] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 614.756329] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] for vif in network_info: [ 614.756329] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 614.756329] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] return self._sync_wrapper(fn, *args, **kwargs) [ 614.756329] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 614.756329] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] self.wait() [ 614.756329] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 614.756329] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] self[:] = self._gt.wait() [ 614.756329] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 614.756329] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] return self._exit_event.wait() [ 614.756329] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 614.756746] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] result = hub.switch() [ 614.756746] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 614.756746] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] return self.greenlet.switch() [ 614.756746] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.756746] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] result = function(*args, **kwargs) [ 614.756746] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.756746] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] return func(*args, **kwargs) [ 614.756746] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.756746] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] raise e [ 614.756746] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.756746] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] nwinfo = self.network_api.allocate_for_instance( [ 614.756746] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 614.756746] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] created_port_ids = self._update_ports_for_instance( [ 614.757174] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 614.757174] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] with excutils.save_and_reraise_exception(): [ 614.757174] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.757174] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] self.force_reraise() [ 614.757174] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.757174] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] raise self.value [ 614.757174] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 614.757174] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] updated_port = self._update_port( [ 614.757174] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.757174] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] _ensure_no_port_binding_failure(port) [ 614.757174] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.757174] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] raise exception.PortBindingFailed(port_id=port['id']) [ 614.757559] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] nova.exception.PortBindingFailed: Binding failed for port 81ec8210-343f-40b7-b92e-ad5a81d2b07e, please check neutron logs for more information. [ 614.757559] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] [ 614.757559] env[62600]: INFO nova.compute.manager [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Terminating instance [ 614.757711] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Acquiring lock "refresh_cache-a934e86f-1a51-468e-915f-add473fc2bf6" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.757876] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Acquired lock "refresh_cache-a934e86f-1a51-468e-915f-add473fc2bf6" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.758064] env[62600]: DEBUG nova.network.neutron [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 614.885671] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': task-1222232, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.444029} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 614.888078] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 8c0e3a15-0cdd-422a-8a97-95d57621de88/8c0e3a15-0cdd-422a-8a97-95d57621de88.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 614.888340] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 614.889415] env[62600]: DEBUG nova.compute.manager [req-7e054563-8ee9-40af-9f3d-9d9717a261b4 req-f3355333-e706-4cb2-8c03-8b6f3264073b service nova] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Received event network-changed-81ec8210-343f-40b7-b92e-ad5a81d2b07e {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 614.889569] env[62600]: DEBUG nova.compute.manager [req-7e054563-8ee9-40af-9f3d-9d9717a261b4 req-f3355333-e706-4cb2-8c03-8b6f3264073b service nova] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Refreshing instance network info cache due to event network-changed-81ec8210-343f-40b7-b92e-ad5a81d2b07e. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 614.889755] env[62600]: DEBUG oslo_concurrency.lockutils [req-7e054563-8ee9-40af-9f3d-9d9717a261b4 req-f3355333-e706-4cb2-8c03-8b6f3264073b service nova] Acquiring lock "refresh_cache-a934e86f-1a51-468e-915f-add473fc2bf6" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.892248] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0d6382fa-723d-4762-823a-0b759bf4d801 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.899354] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Waiting for the task: (returnval){ [ 614.899354] env[62600]: value = "task-1222233" [ 614.899354] env[62600]: _type = "Task" [ 614.899354] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 614.912764] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': task-1222233, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.159925] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f6c7bd-873b-4187-80c4-d5e5fdcac84d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.167572] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f105ea04-f32a-450d-98e8-69cd3891a972 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.201530] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d07e400-241c-40e3-9365-b7a2f0c36202 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.208555] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf04b9ca-da84-47b3-af39-661321aed50a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.221771] env[62600]: DEBUG nova.compute.provider_tree [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.279579] env[62600]: DEBUG nova.network.neutron [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.363114] env[62600]: DEBUG nova.network.neutron [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.399252] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Acquiring lock "014899cb-db4c-4444-abe1-b0129c0d7db9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.401223] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Lock "014899cb-db4c-4444-abe1-b0129c0d7db9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.418016] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': task-1222233, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070061} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.418016] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 615.418016] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a63f81-7858-43a8-8192-8cd101311cf6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.437429] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Reconfiguring VM instance instance-0000000d to attach disk [datastore2] 8c0e3a15-0cdd-422a-8a97-95d57621de88/8c0e3a15-0cdd-422a-8a97-95d57621de88.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 615.438010] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b5cc958-4236-4f3f-9ba9-8c4b46fcf97a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.459384] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Waiting for the task: (returnval){ [ 615.459384] env[62600]: value = "task-1222235" [ 615.459384] env[62600]: _type = "Task" [ 615.459384] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.467037] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': task-1222235, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.725151] env[62600]: DEBUG nova.scheduler.client.report [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 615.865896] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Releasing lock "refresh_cache-a934e86f-1a51-468e-915f-add473fc2bf6" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.865896] env[62600]: DEBUG nova.compute.manager [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 615.866087] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 615.866691] env[62600]: DEBUG oslo_concurrency.lockutils [req-7e054563-8ee9-40af-9f3d-9d9717a261b4 req-f3355333-e706-4cb2-8c03-8b6f3264073b service nova] Acquired lock "refresh_cache-a934e86f-1a51-468e-915f-add473fc2bf6" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.866691] env[62600]: DEBUG nova.network.neutron [req-7e054563-8ee9-40af-9f3d-9d9717a261b4 req-f3355333-e706-4cb2-8c03-8b6f3264073b service nova] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Refreshing network info cache for port 81ec8210-343f-40b7-b92e-ad5a81d2b07e {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 615.867599] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c30f35f-03f0-4300-b047-4fcc64b41e80 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.877343] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b51ab7ee-3d73-4971-b915-654118e461d8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.901885] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a934e86f-1a51-468e-915f-add473fc2bf6 could not be found. [ 615.902114] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 615.902295] env[62600]: INFO nova.compute.manager [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 615.902530] env[62600]: DEBUG oslo.service.loopingcall [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 615.902737] env[62600]: DEBUG nova.compute.manager [-] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.902830] env[62600]: DEBUG nova.network.neutron [-] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 615.918651] env[62600]: DEBUG nova.network.neutron [-] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.968983] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': task-1222235, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.233024] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.597s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.233024] env[62600]: DEBUG nova.compute.manager [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 616.235334] env[62600]: DEBUG oslo_concurrency.lockutils [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.913s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.235588] env[62600]: DEBUG nova.objects.instance [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Lazy-loading 'resources' on Instance uuid 98cbcc28-54e0-40de-9feb-6da917806423 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 616.388128] env[62600]: DEBUG nova.network.neutron [req-7e054563-8ee9-40af-9f3d-9d9717a261b4 req-f3355333-e706-4cb2-8c03-8b6f3264073b service nova] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.422086] env[62600]: DEBUG nova.network.neutron [-] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.464374] env[62600]: DEBUG nova.network.neutron [req-7e054563-8ee9-40af-9f3d-9d9717a261b4 req-f3355333-e706-4cb2-8c03-8b6f3264073b service nova] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.471577] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': task-1222235, 'name': ReconfigVM_Task, 'duration_secs': 0.982366} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.471865] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Reconfigured VM instance instance-0000000d to attach disk [datastore2] 8c0e3a15-0cdd-422a-8a97-95d57621de88/8c0e3a15-0cdd-422a-8a97-95d57621de88.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 616.472515] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc6be061-4ac8-4add-bcd6-5aa438dca70d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.479057] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Waiting for the task: (returnval){ [ 616.479057] env[62600]: value = "task-1222236" [ 616.479057] env[62600]: _type = "Task" [ 616.479057] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.486802] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': task-1222236, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.742040] env[62600]: DEBUG nova.compute.utils [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 616.744176] env[62600]: DEBUG nova.compute.manager [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 616.744331] env[62600]: DEBUG nova.network.neutron [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 616.784672] env[62600]: DEBUG nova.policy [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ecef878e2d494b47b76eeeee20abe886', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'abc5d9b5781247a69855b11a1c71aed2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 616.925705] env[62600]: INFO nova.compute.manager [-] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Took 1.02 seconds to deallocate network for instance. [ 616.929125] env[62600]: DEBUG nova.compute.manager [req-345587fe-551d-4c57-83f7-ce0f5901b0bd req-7f350188-5191-4eae-9a3a-b5c270da16a0 service nova] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Received event network-vif-deleted-81ec8210-343f-40b7-b92e-ad5a81d2b07e {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 616.931028] env[62600]: DEBUG nova.compute.claims [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 616.931028] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.969287] env[62600]: DEBUG oslo_concurrency.lockutils [req-7e054563-8ee9-40af-9f3d-9d9717a261b4 req-f3355333-e706-4cb2-8c03-8b6f3264073b service nova] Releasing lock "refresh_cache-a934e86f-1a51-468e-915f-add473fc2bf6" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.990377] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': task-1222236, 'name': Rename_Task, 'duration_secs': 0.120762} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 616.990835] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 616.991258] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-366ce641-feb5-4c54-9d63-43dba6f71417 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.000623] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Waiting for the task: (returnval){ [ 617.000623] env[62600]: value = "task-1222237" [ 617.000623] env[62600]: _type = "Task" [ 617.000623] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.008281] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': task-1222237, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.037446] env[62600]: DEBUG nova.network.neutron [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Successfully created port: 3859d35c-4111-42ef-9ac2-15f4f99f0f49 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 617.161919] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5118627-2ca9-4a32-899e-c62c07fe3272 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.169666] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80fdef0e-d1e4-43a6-b226-2209816a15b4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.202274] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6968dd9-029e-454b-a311-9391c62f587d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.209983] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2d0a319-9bf2-44f3-9421-5abeb31c05bf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.224121] env[62600]: DEBUG nova.compute.provider_tree [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 617.244765] env[62600]: DEBUG nova.compute.manager [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 617.509575] env[62600]: DEBUG oslo_vmware.api [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': task-1222237, 'name': PowerOnVM_Task, 'duration_secs': 0.42659} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.509856] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 617.510089] env[62600]: INFO nova.compute.manager [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Took 5.41 seconds to spawn the instance on the hypervisor. [ 617.510291] env[62600]: DEBUG nova.compute.manager [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 617.511039] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421bedaf-6dd8-480f-8925-6fb7aed8683c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.734559] env[62600]: DEBUG nova.scheduler.client.report [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 617.937151] env[62600]: ERROR nova.compute.manager [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3859d35c-4111-42ef-9ac2-15f4f99f0f49, please check neutron logs for more information. [ 617.937151] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 617.937151] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.937151] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 617.937151] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 617.937151] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 617.937151] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 617.937151] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 617.937151] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.937151] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 617.937151] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.937151] env[62600]: ERROR nova.compute.manager raise self.value [ 617.937151] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 617.937151] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 617.937151] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.937151] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 617.937702] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.937702] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 617.937702] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3859d35c-4111-42ef-9ac2-15f4f99f0f49, please check neutron logs for more information. [ 617.937702] env[62600]: ERROR nova.compute.manager [ 617.937702] env[62600]: Traceback (most recent call last): [ 617.937702] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 617.937702] env[62600]: listener.cb(fileno) [ 617.937702] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.937702] env[62600]: result = function(*args, **kwargs) [ 617.937702] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 617.937702] env[62600]: return func(*args, **kwargs) [ 617.937702] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.937702] env[62600]: raise e [ 617.937702] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.937702] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 617.937702] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 617.937702] env[62600]: created_port_ids = self._update_ports_for_instance( [ 617.937702] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 617.937702] env[62600]: with excutils.save_and_reraise_exception(): [ 617.937702] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.937702] env[62600]: self.force_reraise() [ 617.937702] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.937702] env[62600]: raise self.value [ 617.937702] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 617.937702] env[62600]: updated_port = self._update_port( [ 617.937702] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.937702] env[62600]: _ensure_no_port_binding_failure(port) [ 617.937702] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.937702] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 617.938661] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 3859d35c-4111-42ef-9ac2-15f4f99f0f49, please check neutron logs for more information. [ 617.938661] env[62600]: Removing descriptor: 16 [ 618.030809] env[62600]: INFO nova.compute.manager [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Took 37.58 seconds to build instance. [ 618.238628] env[62600]: DEBUG oslo_concurrency.lockutils [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.003s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.241873] env[62600]: DEBUG oslo_concurrency.lockutils [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.788s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.242536] env[62600]: INFO nova.compute.claims [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 618.254964] env[62600]: DEBUG nova.compute.manager [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 618.265284] env[62600]: INFO nova.scheduler.client.report [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Deleted allocations for instance 98cbcc28-54e0-40de-9feb-6da917806423 [ 618.289582] env[62600]: DEBUG nova.virt.hardware [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 618.290110] env[62600]: DEBUG nova.virt.hardware [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 618.290110] env[62600]: DEBUG nova.virt.hardware [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 618.290213] env[62600]: DEBUG nova.virt.hardware [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 618.291026] env[62600]: DEBUG nova.virt.hardware [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 618.291026] env[62600]: DEBUG nova.virt.hardware [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 618.291026] env[62600]: DEBUG nova.virt.hardware [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 618.291026] env[62600]: DEBUG nova.virt.hardware [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 618.291026] env[62600]: DEBUG nova.virt.hardware [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 618.291315] env[62600]: DEBUG nova.virt.hardware [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 618.291315] env[62600]: DEBUG nova.virt.hardware [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 618.292199] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2bb66cf-c008-4dc9-82ed-d0b6b3ce57ed {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.300627] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a58ec13c-8e03-4174-bd9f-299e2856c077 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.320559] env[62600]: ERROR nova.compute.manager [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3859d35c-4111-42ef-9ac2-15f4f99f0f49, please check neutron logs for more information. [ 618.320559] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Traceback (most recent call last): [ 618.320559] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 618.320559] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] yield resources [ 618.320559] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.320559] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] self.driver.spawn(context, instance, image_meta, [ 618.320559] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 618.320559] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.320559] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.320559] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] vm_ref = self.build_virtual_machine(instance, [ 618.320559] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.321020] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.321020] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.321020] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] for vif in network_info: [ 618.321020] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.321020] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] return self._sync_wrapper(fn, *args, **kwargs) [ 618.321020] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.321020] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] self.wait() [ 618.321020] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.321020] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] self[:] = self._gt.wait() [ 618.321020] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.321020] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] return self._exit_event.wait() [ 618.321020] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 618.321020] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] current.throw(*self._exc) [ 618.321519] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.321519] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] result = function(*args, **kwargs) [ 618.321519] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.321519] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] return func(*args, **kwargs) [ 618.321519] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.321519] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] raise e [ 618.321519] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.321519] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] nwinfo = self.network_api.allocate_for_instance( [ 618.321519] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 618.321519] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] created_port_ids = self._update_ports_for_instance( [ 618.321519] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 618.321519] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] with excutils.save_and_reraise_exception(): [ 618.321519] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.321977] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] self.force_reraise() [ 618.321977] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.321977] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] raise self.value [ 618.321977] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 618.321977] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] updated_port = self._update_port( [ 618.321977] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.321977] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] _ensure_no_port_binding_failure(port) [ 618.321977] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.321977] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] raise exception.PortBindingFailed(port_id=port['id']) [ 618.321977] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] nova.exception.PortBindingFailed: Binding failed for port 3859d35c-4111-42ef-9ac2-15f4f99f0f49, please check neutron logs for more information. [ 618.321977] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] [ 618.321977] env[62600]: INFO nova.compute.manager [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Terminating instance [ 618.323039] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Acquiring lock "refresh_cache-b3cc82aa-be39-4e89-8972-b5596629961c" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.323170] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Acquired lock "refresh_cache-b3cc82aa-be39-4e89-8972-b5596629961c" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.323332] env[62600]: DEBUG nova.network.neutron [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 618.533824] env[62600]: DEBUG oslo_concurrency.lockutils [None req-647cff55-0ba8-431d-b9d4-d40741589faa tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Lock "8c0e3a15-0cdd-422a-8a97-95d57621de88" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.094s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.581707] env[62600]: DEBUG nova.compute.manager [None req-92d20fd4-10cc-4eaf-97aa-785cb977fa5a tempest-ServerDiagnosticsV248Test-1010709256 tempest-ServerDiagnosticsV248Test-1010709256-project-admin] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 618.582870] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f065ba2-794c-4352-841d-c8ba0f5e79be {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.589981] env[62600]: INFO nova.compute.manager [None req-92d20fd4-10cc-4eaf-97aa-785cb977fa5a tempest-ServerDiagnosticsV248Test-1010709256 tempest-ServerDiagnosticsV248Test-1010709256-project-admin] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Retrieving diagnostics [ 618.590931] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3713e60e-c65b-494a-8a11-93df206b20e6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.775611] env[62600]: DEBUG oslo_concurrency.lockutils [None req-488b424c-e6f2-469e-9945-1ab44ee30a8f tempest-ServersAdmin275Test-858531567 tempest-ServersAdmin275Test-858531567-project-member] Lock "98cbcc28-54e0-40de-9feb-6da917806423" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.302s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.841185] env[62600]: DEBUG nova.network.neutron [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.925232] env[62600]: DEBUG nova.network.neutron [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.951318] env[62600]: DEBUG nova.compute.manager [req-1d7e7058-e942-4dfb-a60c-7976d09b9fec req-a50ddc14-ac66-4d40-9f86-7becbd8c2c35 service nova] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Received event network-changed-3859d35c-4111-42ef-9ac2-15f4f99f0f49 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 618.951391] env[62600]: DEBUG nova.compute.manager [req-1d7e7058-e942-4dfb-a60c-7976d09b9fec req-a50ddc14-ac66-4d40-9f86-7becbd8c2c35 service nova] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Refreshing instance network info cache due to event network-changed-3859d35c-4111-42ef-9ac2-15f4f99f0f49. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 618.951540] env[62600]: DEBUG oslo_concurrency.lockutils [req-1d7e7058-e942-4dfb-a60c-7976d09b9fec req-a50ddc14-ac66-4d40-9f86-7becbd8c2c35 service nova] Acquiring lock "refresh_cache-b3cc82aa-be39-4e89-8972-b5596629961c" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.036864] env[62600]: DEBUG nova.compute.manager [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 619.428355] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Releasing lock "refresh_cache-b3cc82aa-be39-4e89-8972-b5596629961c" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.428816] env[62600]: DEBUG nova.compute.manager [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 619.429019] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 619.429528] env[62600]: DEBUG oslo_concurrency.lockutils [req-1d7e7058-e942-4dfb-a60c-7976d09b9fec req-a50ddc14-ac66-4d40-9f86-7becbd8c2c35 service nova] Acquired lock "refresh_cache-b3cc82aa-be39-4e89-8972-b5596629961c" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.429713] env[62600]: DEBUG nova.network.neutron [req-1d7e7058-e942-4dfb-a60c-7976d09b9fec req-a50ddc14-ac66-4d40-9f86-7becbd8c2c35 service nova] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Refreshing network info cache for port 3859d35c-4111-42ef-9ac2-15f4f99f0f49 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 619.430737] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1495cb45-01dd-4424-bfbe-c0e44b4a286e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.444166] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e069a49-a00a-4e23-a50f-fbe37fcb97c1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.472229] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b3cc82aa-be39-4e89-8972-b5596629961c could not be found. [ 619.472456] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 619.472635] env[62600]: INFO nova.compute.manager [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 619.472955] env[62600]: DEBUG oslo.service.loopingcall [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 619.473374] env[62600]: DEBUG nova.compute.manager [-] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 619.473478] env[62600]: DEBUG nova.network.neutron [-] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 619.490160] env[62600]: DEBUG nova.network.neutron [-] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.563859] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.698259] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e9d5d2c-a209-4825-9f0d-f1a50e4cde81 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.705310] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92773873-7b6a-4cca-902d-aa9725fcfd1f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.736117] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb7815ff-d406-4b0a-ac37-def8ee057c00 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.744070] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83500ef2-cc1f-4e80-9893-96677645ac4d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.758499] env[62600]: DEBUG nova.compute.provider_tree [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 619.956029] env[62600]: DEBUG nova.network.neutron [req-1d7e7058-e942-4dfb-a60c-7976d09b9fec req-a50ddc14-ac66-4d40-9f86-7becbd8c2c35 service nova] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.995250] env[62600]: DEBUG nova.network.neutron [-] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.049134] env[62600]: DEBUG nova.network.neutron [req-1d7e7058-e942-4dfb-a60c-7976d09b9fec req-a50ddc14-ac66-4d40-9f86-7becbd8c2c35 service nova] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.262013] env[62600]: DEBUG nova.scheduler.client.report [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 620.502465] env[62600]: INFO nova.compute.manager [-] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Took 1.03 seconds to deallocate network for instance. [ 620.505085] env[62600]: DEBUG nova.compute.claims [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 620.506980] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.552314] env[62600]: DEBUG oslo_concurrency.lockutils [req-1d7e7058-e942-4dfb-a60c-7976d09b9fec req-a50ddc14-ac66-4d40-9f86-7becbd8c2c35 service nova] Releasing lock "refresh_cache-b3cc82aa-be39-4e89-8972-b5596629961c" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.553152] env[62600]: DEBUG nova.compute.manager [req-1d7e7058-e942-4dfb-a60c-7976d09b9fec req-a50ddc14-ac66-4d40-9f86-7becbd8c2c35 service nova] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Received event network-vif-deleted-3859d35c-4111-42ef-9ac2-15f4f99f0f49 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 620.767025] env[62600]: DEBUG oslo_concurrency.lockutils [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.526s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.767559] env[62600]: DEBUG nova.compute.manager [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 620.770252] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.846s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.275481] env[62600]: DEBUG nova.compute.utils [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 621.280759] env[62600]: DEBUG nova.compute.manager [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 621.280989] env[62600]: DEBUG nova.network.neutron [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 621.343462] env[62600]: DEBUG nova.policy [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ee6ba4a734654b35affeb667847deb62', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0f28c3a28717483c94c1940d2f0d1647', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 621.729063] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7efafd3e-5d4d-4e55-be34-2e1bb8f36423 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.738839] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64bd8b38-9271-42e4-8a30-be4f141843d4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.772889] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edffff31-bf0b-48af-ad03-5ca82fc4e47b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.781298] env[62600]: DEBUG nova.compute.manager [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 621.787842] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41cced1d-84d8-478d-85d5-27a14f858d4f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.804873] env[62600]: DEBUG nova.compute.provider_tree [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.843510] env[62600]: DEBUG nova.network.neutron [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Successfully created port: 88b7061a-9a5f-4724-a991-33a74ab51ae7 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 622.309293] env[62600]: DEBUG nova.scheduler.client.report [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 622.636907] env[62600]: DEBUG nova.compute.manager [req-4014b1ce-e586-4938-a017-9e6adcaec925 req-16d7ca03-2bf1-4b6b-972c-6f3186fb576c service nova] [instance: ba442811-2b0a-44e7-b036-572291aff648] Received event network-changed-88b7061a-9a5f-4724-a991-33a74ab51ae7 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 622.637217] env[62600]: DEBUG nova.compute.manager [req-4014b1ce-e586-4938-a017-9e6adcaec925 req-16d7ca03-2bf1-4b6b-972c-6f3186fb576c service nova] [instance: ba442811-2b0a-44e7-b036-572291aff648] Refreshing instance network info cache due to event network-changed-88b7061a-9a5f-4724-a991-33a74ab51ae7. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 622.637362] env[62600]: DEBUG oslo_concurrency.lockutils [req-4014b1ce-e586-4938-a017-9e6adcaec925 req-16d7ca03-2bf1-4b6b-972c-6f3186fb576c service nova] Acquiring lock "refresh_cache-ba442811-2b0a-44e7-b036-572291aff648" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.637492] env[62600]: DEBUG oslo_concurrency.lockutils [req-4014b1ce-e586-4938-a017-9e6adcaec925 req-16d7ca03-2bf1-4b6b-972c-6f3186fb576c service nova] Acquired lock "refresh_cache-ba442811-2b0a-44e7-b036-572291aff648" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.637948] env[62600]: DEBUG nova.network.neutron [req-4014b1ce-e586-4938-a017-9e6adcaec925 req-16d7ca03-2bf1-4b6b-972c-6f3186fb576c service nova] [instance: ba442811-2b0a-44e7-b036-572291aff648] Refreshing network info cache for port 88b7061a-9a5f-4724-a991-33a74ab51ae7 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 622.799140] env[62600]: DEBUG nova.compute.manager [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 622.806256] env[62600]: ERROR nova.compute.manager [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 88b7061a-9a5f-4724-a991-33a74ab51ae7, please check neutron logs for more information. [ 622.806256] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 622.806256] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.806256] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 622.806256] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 622.806256] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 622.806256] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 622.806256] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 622.806256] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.806256] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 622.806256] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.806256] env[62600]: ERROR nova.compute.manager raise self.value [ 622.806256] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 622.806256] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 622.806256] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.806256] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 622.807488] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.807488] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 622.807488] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 88b7061a-9a5f-4724-a991-33a74ab51ae7, please check neutron logs for more information. [ 622.807488] env[62600]: ERROR nova.compute.manager [ 622.807488] env[62600]: Traceback (most recent call last): [ 622.807488] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 622.807488] env[62600]: listener.cb(fileno) [ 622.807488] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.807488] env[62600]: result = function(*args, **kwargs) [ 622.807488] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.807488] env[62600]: return func(*args, **kwargs) [ 622.807488] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 622.807488] env[62600]: raise e [ 622.807488] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.807488] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 622.807488] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 622.807488] env[62600]: created_port_ids = self._update_ports_for_instance( [ 622.807488] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 622.807488] env[62600]: with excutils.save_and_reraise_exception(): [ 622.807488] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.807488] env[62600]: self.force_reraise() [ 622.807488] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.807488] env[62600]: raise self.value [ 622.807488] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 622.807488] env[62600]: updated_port = self._update_port( [ 622.807488] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.807488] env[62600]: _ensure_no_port_binding_failure(port) [ 622.807488] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.807488] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 622.808461] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 88b7061a-9a5f-4724-a991-33a74ab51ae7, please check neutron logs for more information. [ 622.808461] env[62600]: Removing descriptor: 16 [ 622.812402] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.042s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.812991] env[62600]: ERROR nova.compute.manager [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6b748657-b67e-4eb0-86a8-6a8edf4ee8b9, please check neutron logs for more information. [ 622.812991] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Traceback (most recent call last): [ 622.812991] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 622.812991] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] self.driver.spawn(context, instance, image_meta, [ 622.812991] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 622.812991] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] self._vmops.spawn(context, instance, image_meta, injected_files, [ 622.812991] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 622.812991] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] vm_ref = self.build_virtual_machine(instance, [ 622.812991] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 622.812991] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] vif_infos = vmwarevif.get_vif_info(self._session, [ 622.812991] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 622.814809] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] for vif in network_info: [ 622.814809] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 622.814809] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] return self._sync_wrapper(fn, *args, **kwargs) [ 622.814809] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 622.814809] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] self.wait() [ 622.814809] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 622.814809] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] self[:] = self._gt.wait() [ 622.814809] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 622.814809] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] return self._exit_event.wait() [ 622.814809] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 622.814809] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] result = hub.switch() [ 622.814809] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 622.814809] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] return self.greenlet.switch() [ 622.815231] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.815231] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] result = function(*args, **kwargs) [ 622.815231] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.815231] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] return func(*args, **kwargs) [ 622.815231] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 622.815231] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] raise e [ 622.815231] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.815231] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] nwinfo = self.network_api.allocate_for_instance( [ 622.815231] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 622.815231] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] created_port_ids = self._update_ports_for_instance( [ 622.815231] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 622.815231] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] with excutils.save_and_reraise_exception(): [ 622.815231] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.815695] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] self.force_reraise() [ 622.815695] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.815695] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] raise self.value [ 622.815695] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 622.815695] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] updated_port = self._update_port( [ 622.815695] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.815695] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] _ensure_no_port_binding_failure(port) [ 622.815695] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.815695] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] raise exception.PortBindingFailed(port_id=port['id']) [ 622.815695] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] nova.exception.PortBindingFailed: Binding failed for port 6b748657-b67e-4eb0-86a8-6a8edf4ee8b9, please check neutron logs for more information. [ 622.815695] env[62600]: ERROR nova.compute.manager [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] [ 622.816107] env[62600]: DEBUG nova.compute.utils [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Binding failed for port 6b748657-b67e-4eb0-86a8-6a8edf4ee8b9, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 622.816107] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.415s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.816574] env[62600]: INFO nova.compute.claims [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 622.819316] env[62600]: DEBUG nova.compute.manager [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Build of instance bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621 was re-scheduled: Binding failed for port 6b748657-b67e-4eb0-86a8-6a8edf4ee8b9, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 622.819873] env[62600]: DEBUG nova.compute.manager [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 622.820126] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "refresh_cache-bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.820126] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquired lock "refresh_cache-bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.820299] env[62600]: DEBUG nova.network.neutron [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 622.833909] env[62600]: DEBUG nova.virt.hardware [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 622.834276] env[62600]: DEBUG nova.virt.hardware [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 622.834358] env[62600]: DEBUG nova.virt.hardware [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 622.834537] env[62600]: DEBUG nova.virt.hardware [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 622.834681] env[62600]: DEBUG nova.virt.hardware [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 622.835904] env[62600]: DEBUG nova.virt.hardware [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 622.835904] env[62600]: DEBUG nova.virt.hardware [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 622.835904] env[62600]: DEBUG nova.virt.hardware [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 622.836118] env[62600]: DEBUG nova.virt.hardware [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 622.836166] env[62600]: DEBUG nova.virt.hardware [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 622.836331] env[62600]: DEBUG nova.virt.hardware [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 622.837231] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f2ae69b-6449-4b2b-809d-dae4f14c6395 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.848013] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86633065-ce20-4084-b9bb-6e045a5a5156 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.864330] env[62600]: ERROR nova.compute.manager [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 88b7061a-9a5f-4724-a991-33a74ab51ae7, please check neutron logs for more information. [ 622.864330] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] Traceback (most recent call last): [ 622.864330] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 622.864330] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] yield resources [ 622.864330] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 622.864330] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] self.driver.spawn(context, instance, image_meta, [ 622.864330] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 622.864330] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] self._vmops.spawn(context, instance, image_meta, injected_files, [ 622.864330] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 622.864330] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] vm_ref = self.build_virtual_machine(instance, [ 622.864330] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 622.865130] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] vif_infos = vmwarevif.get_vif_info(self._session, [ 622.865130] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 622.865130] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] for vif in network_info: [ 622.865130] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 622.865130] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] return self._sync_wrapper(fn, *args, **kwargs) [ 622.865130] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 622.865130] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] self.wait() [ 622.865130] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 622.865130] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] self[:] = self._gt.wait() [ 622.865130] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 622.865130] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] return self._exit_event.wait() [ 622.865130] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 622.865130] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] current.throw(*self._exc) [ 622.865655] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.865655] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] result = function(*args, **kwargs) [ 622.865655] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.865655] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] return func(*args, **kwargs) [ 622.865655] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 622.865655] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] raise e [ 622.865655] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.865655] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] nwinfo = self.network_api.allocate_for_instance( [ 622.865655] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 622.865655] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] created_port_ids = self._update_ports_for_instance( [ 622.865655] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 622.865655] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] with excutils.save_and_reraise_exception(): [ 622.865655] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.866138] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] self.force_reraise() [ 622.866138] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.866138] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] raise self.value [ 622.866138] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 622.866138] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] updated_port = self._update_port( [ 622.866138] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.866138] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] _ensure_no_port_binding_failure(port) [ 622.866138] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.866138] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] raise exception.PortBindingFailed(port_id=port['id']) [ 622.866138] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] nova.exception.PortBindingFailed: Binding failed for port 88b7061a-9a5f-4724-a991-33a74ab51ae7, please check neutron logs for more information. [ 622.866138] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] [ 622.866138] env[62600]: INFO nova.compute.manager [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Terminating instance [ 622.866668] env[62600]: DEBUG oslo_concurrency.lockutils [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Acquiring lock "refresh_cache-ba442811-2b0a-44e7-b036-572291aff648" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.155077] env[62600]: DEBUG nova.network.neutron [req-4014b1ce-e586-4938-a017-9e6adcaec925 req-16d7ca03-2bf1-4b6b-972c-6f3186fb576c service nova] [instance: ba442811-2b0a-44e7-b036-572291aff648] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 623.242693] env[62600]: DEBUG nova.network.neutron [req-4014b1ce-e586-4938-a017-9e6adcaec925 req-16d7ca03-2bf1-4b6b-972c-6f3186fb576c service nova] [instance: ba442811-2b0a-44e7-b036-572291aff648] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.343306] env[62600]: DEBUG nova.network.neutron [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 623.387909] env[62600]: DEBUG nova.network.neutron [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.745708] env[62600]: DEBUG oslo_concurrency.lockutils [req-4014b1ce-e586-4938-a017-9e6adcaec925 req-16d7ca03-2bf1-4b6b-972c-6f3186fb576c service nova] Releasing lock "refresh_cache-ba442811-2b0a-44e7-b036-572291aff648" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.746205] env[62600]: DEBUG oslo_concurrency.lockutils [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Acquired lock "refresh_cache-ba442811-2b0a-44e7-b036-572291aff648" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.746526] env[62600]: DEBUG nova.network.neutron [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 623.891705] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Releasing lock "refresh_cache-bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.894929] env[62600]: DEBUG nova.compute.manager [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 623.894929] env[62600]: DEBUG nova.compute.manager [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 623.894929] env[62600]: DEBUG nova.network.neutron [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 623.915708] env[62600]: DEBUG nova.network.neutron [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.267344] env[62600]: DEBUG nova.network.neutron [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.303424] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b76688-6e53-4ffd-8841-fc00c44a9ce4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.313229] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fde42548-8ffe-49f6-adc3-39650eb2dc82 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.344611] env[62600]: DEBUG nova.network.neutron [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.346183] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-796e4e91-7c34-459c-865c-d0fbd8aa049a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.354441] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49905e71-a6ab-490a-bf5b-a2794a293e10 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.368251] env[62600]: DEBUG nova.compute.provider_tree [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.418450] env[62600]: DEBUG nova.network.neutron [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.701899] env[62600]: DEBUG nova.compute.manager [req-b76d8bb8-03bd-44c4-8833-25fc10650c76 req-3d76a76d-3db6-4a1f-87a0-1b54c1653c76 service nova] [instance: ba442811-2b0a-44e7-b036-572291aff648] Received event network-vif-deleted-88b7061a-9a5f-4724-a991-33a74ab51ae7 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 624.849863] env[62600]: DEBUG oslo_concurrency.lockutils [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Releasing lock "refresh_cache-ba442811-2b0a-44e7-b036-572291aff648" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.850350] env[62600]: DEBUG nova.compute.manager [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 624.850544] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 624.851140] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2d548887-c3d7-4863-999f-47ceb5909f04 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.860906] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96560731-e3b3-44e9-99c5-511471e6059d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.872106] env[62600]: DEBUG nova.scheduler.client.report [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.889076] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ba442811-2b0a-44e7-b036-572291aff648 could not be found. [ 624.889311] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 624.889526] env[62600]: INFO nova.compute.manager [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Took 0.04 seconds to destroy the instance on the hypervisor. [ 624.889728] env[62600]: DEBUG oslo.service.loopingcall [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 624.890559] env[62600]: DEBUG nova.compute.manager [-] [instance: ba442811-2b0a-44e7-b036-572291aff648] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 624.890647] env[62600]: DEBUG nova.network.neutron [-] [instance: ba442811-2b0a-44e7-b036-572291aff648] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 624.911824] env[62600]: DEBUG nova.network.neutron [-] [instance: ba442811-2b0a-44e7-b036-572291aff648] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.921878] env[62600]: INFO nova.compute.manager [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621] Took 1.03 seconds to deallocate network for instance. [ 625.377042] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.562s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.377514] env[62600]: DEBUG nova.compute.manager [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 625.380014] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.963s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.381395] env[62600]: INFO nova.compute.claims [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 625.413688] env[62600]: DEBUG nova.network.neutron [-] [instance: ba442811-2b0a-44e7-b036-572291aff648] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.885396] env[62600]: DEBUG nova.compute.utils [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 625.888676] env[62600]: DEBUG nova.compute.manager [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 625.889792] env[62600]: DEBUG nova.network.neutron [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 625.916673] env[62600]: INFO nova.compute.manager [-] [instance: ba442811-2b0a-44e7-b036-572291aff648] Took 1.03 seconds to deallocate network for instance. [ 625.919020] env[62600]: DEBUG nova.compute.claims [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 625.919213] env[62600]: DEBUG oslo_concurrency.lockutils [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.938829] env[62600]: DEBUG nova.policy [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '04f4af704ccc4e85820f3fa8c7dbb956', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '41ec3023d88247a8ba8b1b418e16c538', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 625.957325] env[62600]: INFO nova.scheduler.client.report [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Deleted allocations for instance bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621 [ 626.196684] env[62600]: DEBUG nova.network.neutron [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Successfully created port: 8871c476-5afb-4342-b4b8-725f0da33941 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 626.389668] env[62600]: DEBUG nova.compute.manager [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 626.467505] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1d7221e6-307a-4019-8808-059d0f6c9913 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "bd1e3e4d-c96d-46d0-89d9-7e5ab6e0f621" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.590s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.807222] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-559ce281-7b52-4a19-8e29-7b9ea53f9ae1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.814193] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e4421f4-19ca-4818-b280-b615bc6b1411 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.849450] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e7e23c-728a-43ea-ab41-1fb37e945554 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.857941] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f186d597-80b5-4e9f-9f9b-f110ab28c258 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.871970] env[62600]: DEBUG nova.compute.provider_tree [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.970608] env[62600]: DEBUG nova.compute.manager [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 626.988630] env[62600]: DEBUG nova.compute.manager [req-6ff3dba0-71bd-4419-a15a-be9200e0e104 req-b7dad75c-ef72-4462-8169-0b71c9222b35 service nova] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Received event network-changed-8871c476-5afb-4342-b4b8-725f0da33941 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 626.988892] env[62600]: DEBUG nova.compute.manager [req-6ff3dba0-71bd-4419-a15a-be9200e0e104 req-b7dad75c-ef72-4462-8169-0b71c9222b35 service nova] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Refreshing instance network info cache due to event network-changed-8871c476-5afb-4342-b4b8-725f0da33941. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 626.989054] env[62600]: DEBUG oslo_concurrency.lockutils [req-6ff3dba0-71bd-4419-a15a-be9200e0e104 req-b7dad75c-ef72-4462-8169-0b71c9222b35 service nova] Acquiring lock "refresh_cache-b569fb64-27c2-4e20-a2c4-c81058958f98" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.989194] env[62600]: DEBUG oslo_concurrency.lockutils [req-6ff3dba0-71bd-4419-a15a-be9200e0e104 req-b7dad75c-ef72-4462-8169-0b71c9222b35 service nova] Acquired lock "refresh_cache-b569fb64-27c2-4e20-a2c4-c81058958f98" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.989356] env[62600]: DEBUG nova.network.neutron [req-6ff3dba0-71bd-4419-a15a-be9200e0e104 req-b7dad75c-ef72-4462-8169-0b71c9222b35 service nova] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Refreshing network info cache for port 8871c476-5afb-4342-b4b8-725f0da33941 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 627.070773] env[62600]: ERROR nova.compute.manager [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8871c476-5afb-4342-b4b8-725f0da33941, please check neutron logs for more information. [ 627.070773] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 627.070773] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.070773] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 627.070773] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 627.070773] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 627.070773] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 627.070773] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 627.070773] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.070773] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 627.070773] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.070773] env[62600]: ERROR nova.compute.manager raise self.value [ 627.070773] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 627.070773] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 627.070773] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.070773] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 627.071453] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.071453] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 627.071453] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8871c476-5afb-4342-b4b8-725f0da33941, please check neutron logs for more information. [ 627.071453] env[62600]: ERROR nova.compute.manager [ 627.071453] env[62600]: Traceback (most recent call last): [ 627.071453] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 627.071453] env[62600]: listener.cb(fileno) [ 627.071453] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.071453] env[62600]: result = function(*args, **kwargs) [ 627.071453] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.071453] env[62600]: return func(*args, **kwargs) [ 627.071453] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.071453] env[62600]: raise e [ 627.071453] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.071453] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 627.071453] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 627.071453] env[62600]: created_port_ids = self._update_ports_for_instance( [ 627.071453] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 627.071453] env[62600]: with excutils.save_and_reraise_exception(): [ 627.071453] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.071453] env[62600]: self.force_reraise() [ 627.071453] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.071453] env[62600]: raise self.value [ 627.071453] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 627.071453] env[62600]: updated_port = self._update_port( [ 627.071453] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.071453] env[62600]: _ensure_no_port_binding_failure(port) [ 627.071453] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.071453] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 627.072483] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 8871c476-5afb-4342-b4b8-725f0da33941, please check neutron logs for more information. [ 627.072483] env[62600]: Removing descriptor: 16 [ 627.377519] env[62600]: DEBUG nova.scheduler.client.report [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 627.402237] env[62600]: DEBUG nova.compute.manager [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 627.426642] env[62600]: DEBUG nova.virt.hardware [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 627.426894] env[62600]: DEBUG nova.virt.hardware [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 627.427066] env[62600]: DEBUG nova.virt.hardware [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 627.427256] env[62600]: DEBUG nova.virt.hardware [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 627.427405] env[62600]: DEBUG nova.virt.hardware [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 627.427551] env[62600]: DEBUG nova.virt.hardware [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 627.427918] env[62600]: DEBUG nova.virt.hardware [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 627.428168] env[62600]: DEBUG nova.virt.hardware [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 627.428358] env[62600]: DEBUG nova.virt.hardware [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 627.428530] env[62600]: DEBUG nova.virt.hardware [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 627.428704] env[62600]: DEBUG nova.virt.hardware [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 627.429868] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443fb221-92c6-4867-97cb-ed61c089595d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.441433] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-681b34f3-c509-4803-b6fd-3808ffb43a73 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.456867] env[62600]: ERROR nova.compute.manager [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8871c476-5afb-4342-b4b8-725f0da33941, please check neutron logs for more information. [ 627.456867] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Traceback (most recent call last): [ 627.456867] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 627.456867] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] yield resources [ 627.456867] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 627.456867] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] self.driver.spawn(context, instance, image_meta, [ 627.456867] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 627.456867] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.456867] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.456867] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] vm_ref = self.build_virtual_machine(instance, [ 627.456867] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.457323] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.457323] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.457323] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] for vif in network_info: [ 627.457323] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.457323] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] return self._sync_wrapper(fn, *args, **kwargs) [ 627.457323] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.457323] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] self.wait() [ 627.457323] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.457323] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] self[:] = self._gt.wait() [ 627.457323] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.457323] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] return self._exit_event.wait() [ 627.457323] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 627.457323] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] current.throw(*self._exc) [ 627.457800] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.457800] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] result = function(*args, **kwargs) [ 627.457800] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.457800] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] return func(*args, **kwargs) [ 627.457800] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.457800] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] raise e [ 627.457800] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.457800] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] nwinfo = self.network_api.allocate_for_instance( [ 627.457800] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 627.457800] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] created_port_ids = self._update_ports_for_instance( [ 627.457800] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 627.457800] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] with excutils.save_and_reraise_exception(): [ 627.457800] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.458234] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] self.force_reraise() [ 627.458234] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.458234] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] raise self.value [ 627.458234] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 627.458234] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] updated_port = self._update_port( [ 627.458234] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.458234] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] _ensure_no_port_binding_failure(port) [ 627.458234] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.458234] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] raise exception.PortBindingFailed(port_id=port['id']) [ 627.458234] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] nova.exception.PortBindingFailed: Binding failed for port 8871c476-5afb-4342-b4b8-725f0da33941, please check neutron logs for more information. [ 627.458234] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] [ 627.458234] env[62600]: INFO nova.compute.manager [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Terminating instance [ 627.459639] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquiring lock "refresh_cache-b569fb64-27c2-4e20-a2c4-c81058958f98" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.491505] env[62600]: DEBUG oslo_concurrency.lockutils [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.508007] env[62600]: DEBUG nova.network.neutron [req-6ff3dba0-71bd-4419-a15a-be9200e0e104 req-b7dad75c-ef72-4462-8169-0b71c9222b35 service nova] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 627.596310] env[62600]: DEBUG nova.network.neutron [req-6ff3dba0-71bd-4419-a15a-be9200e0e104 req-b7dad75c-ef72-4462-8169-0b71c9222b35 service nova] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.883185] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.503s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.884141] env[62600]: DEBUG nova.compute.manager [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 627.886279] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.356s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.887878] env[62600]: INFO nova.compute.claims [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 628.098680] env[62600]: DEBUG oslo_concurrency.lockutils [req-6ff3dba0-71bd-4419-a15a-be9200e0e104 req-b7dad75c-ef72-4462-8169-0b71c9222b35 service nova] Releasing lock "refresh_cache-b569fb64-27c2-4e20-a2c4-c81058958f98" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.099096] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquired lock "refresh_cache-b569fb64-27c2-4e20-a2c4-c81058958f98" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.099331] env[62600]: DEBUG nova.network.neutron [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 628.358831] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "de4f9304-3357-4eaa-9c94-fe28bc554086" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.359078] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "de4f9304-3357-4eaa-9c94-fe28bc554086" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.391770] env[62600]: DEBUG nova.compute.utils [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 628.397935] env[62600]: DEBUG nova.compute.manager [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 628.398149] env[62600]: DEBUG nova.network.neutron [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 628.439265] env[62600]: DEBUG nova.policy [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '04f4af704ccc4e85820f3fa8c7dbb956', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '41ec3023d88247a8ba8b1b418e16c538', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 628.622114] env[62600]: DEBUG nova.network.neutron [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.745812] env[62600]: DEBUG nova.network.neutron [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.759606] env[62600]: DEBUG nova.network.neutron [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Successfully created port: 98868f07-201e-47e6-a446-99bfe9262854 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 628.901938] env[62600]: DEBUG nova.compute.manager [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 629.121562] env[62600]: DEBUG nova.compute.manager [req-fe5b3431-be82-4725-bfa9-6ee03f26ba17 req-951234ba-5057-452c-8a90-33e75ac4dfa9 service nova] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Received event network-vif-deleted-8871c476-5afb-4342-b4b8-725f0da33941 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 629.147115] env[62600]: DEBUG nova.compute.manager [None req-68d7b455-0dfa-41ab-8104-8bc7c1ae8041 tempest-ServerDiagnosticsV248Test-1010709256 tempest-ServerDiagnosticsV248Test-1010709256-project-admin] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 629.150499] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6383a8e0-99a2-4f75-a0c0-edabf9bdf52e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.158471] env[62600]: INFO nova.compute.manager [None req-68d7b455-0dfa-41ab-8104-8bc7c1ae8041 tempest-ServerDiagnosticsV248Test-1010709256 tempest-ServerDiagnosticsV248Test-1010709256-project-admin] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Retrieving diagnostics [ 629.159222] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0774c2f-a203-4f47-81b9-446d6e4a1f66 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.247918] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Releasing lock "refresh_cache-b569fb64-27c2-4e20-a2c4-c81058958f98" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.248345] env[62600]: DEBUG nova.compute.manager [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 629.248534] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 629.248808] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cfb4674a-b2f8-4d31-9d4a-cf087c2d723b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.260299] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3934dd3e-b7b5-4c81-ae06-d08d1efb2f71 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.288500] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b569fb64-27c2-4e20-a2c4-c81058958f98 could not be found. [ 629.288722] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 629.288902] env[62600]: INFO nova.compute.manager [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Took 0.04 seconds to destroy the instance on the hypervisor. [ 629.289168] env[62600]: DEBUG oslo.service.loopingcall [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 629.289396] env[62600]: DEBUG nova.compute.manager [-] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 629.289491] env[62600]: DEBUG nova.network.neutron [-] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 629.309682] env[62600]: DEBUG nova.network.neutron [-] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.340649] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d72ad62b-3a05-490b-80b3-02f5cd2c0555 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.348965] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16fcfc07-810a-4363-b5ca-28db0fee9059 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.379111] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-903e30cc-1ccd-400c-bcd7-d33ab11e66e4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.387385] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4adb7ab5-e28b-4a46-92e3-2a29bdbbe90f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.401139] env[62600]: DEBUG nova.compute.provider_tree [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.722858] env[62600]: ERROR nova.compute.manager [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 98868f07-201e-47e6-a446-99bfe9262854, please check neutron logs for more information. [ 629.722858] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 629.722858] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.722858] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 629.722858] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 629.722858] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 629.722858] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 629.722858] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 629.722858] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.722858] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 629.722858] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.722858] env[62600]: ERROR nova.compute.manager raise self.value [ 629.722858] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 629.722858] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 629.722858] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.722858] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 629.723284] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.723284] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 629.723284] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 98868f07-201e-47e6-a446-99bfe9262854, please check neutron logs for more information. [ 629.723284] env[62600]: ERROR nova.compute.manager [ 629.723284] env[62600]: Traceback (most recent call last): [ 629.723284] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 629.723284] env[62600]: listener.cb(fileno) [ 629.723284] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.723284] env[62600]: result = function(*args, **kwargs) [ 629.723284] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.723284] env[62600]: return func(*args, **kwargs) [ 629.723284] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.723284] env[62600]: raise e [ 629.723284] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.723284] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 629.723284] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 629.723284] env[62600]: created_port_ids = self._update_ports_for_instance( [ 629.723284] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 629.723284] env[62600]: with excutils.save_and_reraise_exception(): [ 629.723284] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.723284] env[62600]: self.force_reraise() [ 629.723284] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.723284] env[62600]: raise self.value [ 629.723284] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 629.723284] env[62600]: updated_port = self._update_port( [ 629.723284] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.723284] env[62600]: _ensure_no_port_binding_failure(port) [ 629.723284] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.723284] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 629.723964] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 98868f07-201e-47e6-a446-99bfe9262854, please check neutron logs for more information. [ 629.723964] env[62600]: Removing descriptor: 16 [ 629.813347] env[62600]: DEBUG nova.network.neutron [-] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.904206] env[62600]: DEBUG nova.scheduler.client.report [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 629.910392] env[62600]: DEBUG nova.compute.manager [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 629.932778] env[62600]: DEBUG nova.virt.hardware [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 629.933077] env[62600]: DEBUG nova.virt.hardware [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 629.933237] env[62600]: DEBUG nova.virt.hardware [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 629.933431] env[62600]: DEBUG nova.virt.hardware [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 629.933598] env[62600]: DEBUG nova.virt.hardware [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 629.933750] env[62600]: DEBUG nova.virt.hardware [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 629.934086] env[62600]: DEBUG nova.virt.hardware [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 629.934258] env[62600]: DEBUG nova.virt.hardware [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 629.934411] env[62600]: DEBUG nova.virt.hardware [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 629.934591] env[62600]: DEBUG nova.virt.hardware [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 629.934765] env[62600]: DEBUG nova.virt.hardware [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 629.935958] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1659cc-3810-41ce-a790-4d50f3402ebc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.946345] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7628e73-4680-4caf-8add-18839f202faf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.960902] env[62600]: ERROR nova.compute.manager [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 98868f07-201e-47e6-a446-99bfe9262854, please check neutron logs for more information. [ 629.960902] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Traceback (most recent call last): [ 629.960902] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 629.960902] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] yield resources [ 629.960902] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 629.960902] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] self.driver.spawn(context, instance, image_meta, [ 629.960902] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 629.960902] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.960902] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.960902] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] vm_ref = self.build_virtual_machine(instance, [ 629.960902] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.961287] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.961287] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.961287] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] for vif in network_info: [ 629.961287] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.961287] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] return self._sync_wrapper(fn, *args, **kwargs) [ 629.961287] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.961287] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] self.wait() [ 629.961287] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.961287] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] self[:] = self._gt.wait() [ 629.961287] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.961287] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] return self._exit_event.wait() [ 629.961287] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 629.961287] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] current.throw(*self._exc) [ 629.961577] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.961577] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] result = function(*args, **kwargs) [ 629.961577] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.961577] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] return func(*args, **kwargs) [ 629.961577] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.961577] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] raise e [ 629.961577] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.961577] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] nwinfo = self.network_api.allocate_for_instance( [ 629.961577] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 629.961577] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] created_port_ids = self._update_ports_for_instance( [ 629.961577] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 629.961577] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] with excutils.save_and_reraise_exception(): [ 629.961577] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.961835] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] self.force_reraise() [ 629.961835] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.961835] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] raise self.value [ 629.961835] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 629.961835] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] updated_port = self._update_port( [ 629.961835] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.961835] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] _ensure_no_port_binding_failure(port) [ 629.961835] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.961835] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] raise exception.PortBindingFailed(port_id=port['id']) [ 629.961835] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] nova.exception.PortBindingFailed: Binding failed for port 98868f07-201e-47e6-a446-99bfe9262854, please check neutron logs for more information. [ 629.961835] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] [ 629.961835] env[62600]: INFO nova.compute.manager [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Terminating instance [ 629.962691] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquiring lock "refresh_cache-dc77594b-7767-40bf-adbb-c72fa5c80503" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.963019] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquired lock "refresh_cache-dc77594b-7767-40bf-adbb-c72fa5c80503" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.963019] env[62600]: DEBUG nova.network.neutron [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 630.305220] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Acquiring lock "8c0e3a15-0cdd-422a-8a97-95d57621de88" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.305551] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Lock "8c0e3a15-0cdd-422a-8a97-95d57621de88" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.305711] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Acquiring lock "8c0e3a15-0cdd-422a-8a97-95d57621de88-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.306349] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Lock "8c0e3a15-0cdd-422a-8a97-95d57621de88-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.306576] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Lock "8c0e3a15-0cdd-422a-8a97-95d57621de88-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.308762] env[62600]: INFO nova.compute.manager [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Terminating instance [ 630.310434] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Acquiring lock "refresh_cache-8c0e3a15-0cdd-422a-8a97-95d57621de88" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.310762] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Acquired lock "refresh_cache-8c0e3a15-0cdd-422a-8a97-95d57621de88" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.310762] env[62600]: DEBUG nova.network.neutron [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 630.315245] env[62600]: INFO nova.compute.manager [-] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Took 1.03 seconds to deallocate network for instance. [ 630.320771] env[62600]: DEBUG nova.compute.claims [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 630.320942] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.408810] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.409400] env[62600]: DEBUG nova.compute.manager [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 630.412103] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.829s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.414024] env[62600]: INFO nova.compute.claims [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 630.481570] env[62600]: DEBUG nova.network.neutron [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.559426] env[62600]: DEBUG nova.network.neutron [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.829647] env[62600]: DEBUG nova.network.neutron [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.879208] env[62600]: DEBUG nova.network.neutron [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.917947] env[62600]: DEBUG nova.compute.utils [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 630.921140] env[62600]: DEBUG nova.compute.manager [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 630.921310] env[62600]: DEBUG nova.network.neutron [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 630.964132] env[62600]: DEBUG nova.policy [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd1715af2650c4d8b9bc0c298079d26d6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ce5ddfe5b316467698a3a047c0074e7f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 631.062228] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Acquiring lock "92cf1acb-b471-4add-a73f-7e9e94fbaaa4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.062722] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Lock "92cf1acb-b471-4add-a73f-7e9e94fbaaa4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.062875] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Releasing lock "refresh_cache-dc77594b-7767-40bf-adbb-c72fa5c80503" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.063279] env[62600]: DEBUG nova.compute.manager [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 631.063535] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 631.064066] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3c6f04f3-4ace-4e07-83e4-d5c27cd4899b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.074144] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b2856c-93a2-4183-80fb-b71bb413eb59 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.099044] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dc77594b-7767-40bf-adbb-c72fa5c80503 could not be found. [ 631.099406] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 631.099584] env[62600]: INFO nova.compute.manager [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Took 0.04 seconds to destroy the instance on the hypervisor. [ 631.099898] env[62600]: DEBUG oslo.service.loopingcall [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 631.100164] env[62600]: DEBUG nova.compute.manager [-] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.100294] env[62600]: DEBUG nova.network.neutron [-] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 631.120734] env[62600]: DEBUG nova.network.neutron [-] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.146992] env[62600]: DEBUG nova.compute.manager [req-a368d8c3-226b-43aa-a3ea-a0862df0def6 req-f95194de-b60a-428d-9e62-10ad1b29397c service nova] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Received event network-changed-98868f07-201e-47e6-a446-99bfe9262854 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 631.148959] env[62600]: DEBUG nova.compute.manager [req-a368d8c3-226b-43aa-a3ea-a0862df0def6 req-f95194de-b60a-428d-9e62-10ad1b29397c service nova] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Refreshing instance network info cache due to event network-changed-98868f07-201e-47e6-a446-99bfe9262854. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 631.149234] env[62600]: DEBUG oslo_concurrency.lockutils [req-a368d8c3-226b-43aa-a3ea-a0862df0def6 req-f95194de-b60a-428d-9e62-10ad1b29397c service nova] Acquiring lock "refresh_cache-dc77594b-7767-40bf-adbb-c72fa5c80503" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.149346] env[62600]: DEBUG oslo_concurrency.lockutils [req-a368d8c3-226b-43aa-a3ea-a0862df0def6 req-f95194de-b60a-428d-9e62-10ad1b29397c service nova] Acquired lock "refresh_cache-dc77594b-7767-40bf-adbb-c72fa5c80503" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.149506] env[62600]: DEBUG nova.network.neutron [req-a368d8c3-226b-43aa-a3ea-a0862df0def6 req-f95194de-b60a-428d-9e62-10ad1b29397c service nova] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Refreshing network info cache for port 98868f07-201e-47e6-a446-99bfe9262854 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 631.260166] env[62600]: DEBUG nova.network.neutron [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Successfully created port: cb7cc583-07cb-4cfd-8f10-8d81c32604f8 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 631.381847] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Releasing lock "refresh_cache-8c0e3a15-0cdd-422a-8a97-95d57621de88" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.382344] env[62600]: DEBUG nova.compute.manager [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 631.382539] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 631.383434] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f28ca49-a44e-4ed5-a533-2367914b4cc2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.393550] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 631.393807] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-70fb7fc1-af39-4538-ad8d-5666e36fba7d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.406717] env[62600]: DEBUG oslo_vmware.api [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Waiting for the task: (returnval){ [ 631.406717] env[62600]: value = "task-1222243" [ 631.406717] env[62600]: _type = "Task" [ 631.406717] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.415799] env[62600]: DEBUG oslo_vmware.api [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': task-1222243, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.422132] env[62600]: DEBUG nova.compute.manager [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 631.624152] env[62600]: DEBUG nova.network.neutron [-] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.671254] env[62600]: DEBUG nova.network.neutron [req-a368d8c3-226b-43aa-a3ea-a0862df0def6 req-f95194de-b60a-428d-9e62-10ad1b29397c service nova] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.792673] env[62600]: DEBUG nova.network.neutron [req-a368d8c3-226b-43aa-a3ea-a0862df0def6 req-f95194de-b60a-428d-9e62-10ad1b29397c service nova] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.917942] env[62600]: DEBUG oslo_vmware.api [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': task-1222243, 'name': PowerOffVM_Task, 'duration_secs': 0.197901} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.918237] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 631.918401] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 631.918657] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9a4a3fa7-d5d3-43fc-81fd-0b94a2a25ac0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.944785] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86bbcfa5-dbe7-47f2-9537-6fc187fdc059 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.957787] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6651fe5-6772-42b0-9805-909e074c15f3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.965083] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 631.965083] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 631.965083] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Deleting the datastore file [datastore2] 8c0e3a15-0cdd-422a-8a97-95d57621de88 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 631.965646] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b6568995-1e9d-455d-acb4-3d3ee68ceccd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.001551] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11b95dcb-a88f-46b6-9693-1ef729cbff50 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.004547] env[62600]: DEBUG oslo_vmware.api [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Waiting for the task: (returnval){ [ 632.004547] env[62600]: value = "task-1222245" [ 632.004547] env[62600]: _type = "Task" [ 632.004547] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.014276] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-847d1aaa-62a7-4636-b305-8d431b919ea5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.022334] env[62600]: DEBUG oslo_vmware.api [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': task-1222245, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.035761] env[62600]: DEBUG nova.compute.provider_tree [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.127059] env[62600]: INFO nova.compute.manager [-] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Took 1.03 seconds to deallocate network for instance. [ 632.130889] env[62600]: DEBUG nova.compute.claims [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 632.131219] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.296021] env[62600]: DEBUG oslo_concurrency.lockutils [req-a368d8c3-226b-43aa-a3ea-a0862df0def6 req-f95194de-b60a-428d-9e62-10ad1b29397c service nova] Releasing lock "refresh_cache-dc77594b-7767-40bf-adbb-c72fa5c80503" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.296021] env[62600]: DEBUG nova.compute.manager [req-a368d8c3-226b-43aa-a3ea-a0862df0def6 req-f95194de-b60a-428d-9e62-10ad1b29397c service nova] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Received event network-vif-deleted-98868f07-201e-47e6-a446-99bfe9262854 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 632.297978] env[62600]: ERROR nova.compute.manager [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cb7cc583-07cb-4cfd-8f10-8d81c32604f8, please check neutron logs for more information. [ 632.297978] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 632.297978] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.297978] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 632.297978] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.297978] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 632.297978] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.297978] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 632.297978] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.297978] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 632.297978] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.297978] env[62600]: ERROR nova.compute.manager raise self.value [ 632.297978] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.297978] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 632.297978] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.297978] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 632.298685] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.298685] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 632.298685] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cb7cc583-07cb-4cfd-8f10-8d81c32604f8, please check neutron logs for more information. [ 632.298685] env[62600]: ERROR nova.compute.manager [ 632.298685] env[62600]: Traceback (most recent call last): [ 632.298685] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 632.298685] env[62600]: listener.cb(fileno) [ 632.298685] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.298685] env[62600]: result = function(*args, **kwargs) [ 632.298685] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.298685] env[62600]: return func(*args, **kwargs) [ 632.298685] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.298685] env[62600]: raise e [ 632.298685] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.298685] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 632.298685] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.298685] env[62600]: created_port_ids = self._update_ports_for_instance( [ 632.298685] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.298685] env[62600]: with excutils.save_and_reraise_exception(): [ 632.298685] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.298685] env[62600]: self.force_reraise() [ 632.298685] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.298685] env[62600]: raise self.value [ 632.298685] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.298685] env[62600]: updated_port = self._update_port( [ 632.298685] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.298685] env[62600]: _ensure_no_port_binding_failure(port) [ 632.298685] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.298685] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 632.299832] env[62600]: nova.exception.PortBindingFailed: Binding failed for port cb7cc583-07cb-4cfd-8f10-8d81c32604f8, please check neutron logs for more information. [ 632.299832] env[62600]: Removing descriptor: 16 [ 632.436776] env[62600]: DEBUG nova.compute.manager [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 632.464595] env[62600]: DEBUG nova.virt.hardware [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 632.464863] env[62600]: DEBUG nova.virt.hardware [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 632.465072] env[62600]: DEBUG nova.virt.hardware [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 632.465273] env[62600]: DEBUG nova.virt.hardware [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 632.465466] env[62600]: DEBUG nova.virt.hardware [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 632.465628] env[62600]: DEBUG nova.virt.hardware [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 632.465824] env[62600]: DEBUG nova.virt.hardware [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 632.465980] env[62600]: DEBUG nova.virt.hardware [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 632.466165] env[62600]: DEBUG nova.virt.hardware [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 632.466737] env[62600]: DEBUG nova.virt.hardware [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 632.466737] env[62600]: DEBUG nova.virt.hardware [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 632.467395] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1480eea1-0b37-4eb6-8e93-14e5ed45bc09 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.476925] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfb219c8-00a8-42d0-869e-ed5a2eb0b50a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.491787] env[62600]: ERROR nova.compute.manager [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cb7cc583-07cb-4cfd-8f10-8d81c32604f8, please check neutron logs for more information. [ 632.491787] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Traceback (most recent call last): [ 632.491787] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 632.491787] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] yield resources [ 632.491787] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 632.491787] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] self.driver.spawn(context, instance, image_meta, [ 632.491787] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 632.491787] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.491787] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.491787] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] vm_ref = self.build_virtual_machine(instance, [ 632.491787] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.492092] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.492092] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.492092] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] for vif in network_info: [ 632.492092] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 632.492092] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] return self._sync_wrapper(fn, *args, **kwargs) [ 632.492092] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 632.492092] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] self.wait() [ 632.492092] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 632.492092] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] self[:] = self._gt.wait() [ 632.492092] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.492092] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] return self._exit_event.wait() [ 632.492092] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 632.492092] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] current.throw(*self._exc) [ 632.492382] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.492382] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] result = function(*args, **kwargs) [ 632.492382] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.492382] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] return func(*args, **kwargs) [ 632.492382] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.492382] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] raise e [ 632.492382] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.492382] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] nwinfo = self.network_api.allocate_for_instance( [ 632.492382] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.492382] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] created_port_ids = self._update_ports_for_instance( [ 632.492382] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.492382] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] with excutils.save_and_reraise_exception(): [ 632.492382] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.492675] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] self.force_reraise() [ 632.492675] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.492675] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] raise self.value [ 632.492675] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.492675] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] updated_port = self._update_port( [ 632.492675] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.492675] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] _ensure_no_port_binding_failure(port) [ 632.492675] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.492675] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] raise exception.PortBindingFailed(port_id=port['id']) [ 632.492675] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] nova.exception.PortBindingFailed: Binding failed for port cb7cc583-07cb-4cfd-8f10-8d81c32604f8, please check neutron logs for more information. [ 632.492675] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] [ 632.492675] env[62600]: INFO nova.compute.manager [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Terminating instance [ 632.494252] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Acquiring lock "refresh_cache-8e64b614-31e3-4830-b620-ff8409da8c37" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.494412] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Acquired lock "refresh_cache-8e64b614-31e3-4830-b620-ff8409da8c37" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.494577] env[62600]: DEBUG nova.network.neutron [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 632.517324] env[62600]: DEBUG oslo_vmware.api [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Task: {'id': task-1222245, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159675} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.517324] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 632.517324] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 632.517324] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 632.517324] env[62600]: INFO nova.compute.manager [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Took 1.13 seconds to destroy the instance on the hypervisor. [ 632.517514] env[62600]: DEBUG oslo.service.loopingcall [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 632.517597] env[62600]: DEBUG nova.compute.manager [-] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.517686] env[62600]: DEBUG nova.network.neutron [-] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 632.535761] env[62600]: DEBUG nova.network.neutron [-] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.538205] env[62600]: DEBUG nova.scheduler.client.report [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 633.013628] env[62600]: DEBUG nova.network.neutron [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.038380] env[62600]: DEBUG nova.network.neutron [-] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.043229] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.631s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.044681] env[62600]: DEBUG nova.compute.manager [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 633.048334] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.836s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.106788] env[62600]: DEBUG nova.network.neutron [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.175072] env[62600]: DEBUG nova.compute.manager [req-ebe8a09c-4d86-424d-a131-e6d62735614b req-87ddce1f-fdd2-4502-aa03-af8171184845 service nova] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Received event network-changed-cb7cc583-07cb-4cfd-8f10-8d81c32604f8 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 633.175279] env[62600]: DEBUG nova.compute.manager [req-ebe8a09c-4d86-424d-a131-e6d62735614b req-87ddce1f-fdd2-4502-aa03-af8171184845 service nova] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Refreshing instance network info cache due to event network-changed-cb7cc583-07cb-4cfd-8f10-8d81c32604f8. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 633.175534] env[62600]: DEBUG oslo_concurrency.lockutils [req-ebe8a09c-4d86-424d-a131-e6d62735614b req-87ddce1f-fdd2-4502-aa03-af8171184845 service nova] Acquiring lock "refresh_cache-8e64b614-31e3-4830-b620-ff8409da8c37" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.541907] env[62600]: INFO nova.compute.manager [-] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Took 1.02 seconds to deallocate network for instance. [ 633.554430] env[62600]: DEBUG nova.compute.utils [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 633.555654] env[62600]: DEBUG nova.compute.manager [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 633.555817] env[62600]: DEBUG nova.network.neutron [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 633.600309] env[62600]: DEBUG nova.policy [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '04f4af704ccc4e85820f3fa8c7dbb956', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '41ec3023d88247a8ba8b1b418e16c538', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 633.610125] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Releasing lock "refresh_cache-8e64b614-31e3-4830-b620-ff8409da8c37" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.610415] env[62600]: DEBUG nova.compute.manager [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 633.610578] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 633.611122] env[62600]: DEBUG oslo_concurrency.lockutils [req-ebe8a09c-4d86-424d-a131-e6d62735614b req-87ddce1f-fdd2-4502-aa03-af8171184845 service nova] Acquired lock "refresh_cache-8e64b614-31e3-4830-b620-ff8409da8c37" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.611303] env[62600]: DEBUG nova.network.neutron [req-ebe8a09c-4d86-424d-a131-e6d62735614b req-87ddce1f-fdd2-4502-aa03-af8171184845 service nova] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Refreshing network info cache for port cb7cc583-07cb-4cfd-8f10-8d81c32604f8 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 633.612280] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4f5d32a7-9bed-45ea-9707-14578753cdbc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.623226] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8bd66cf-ab6e-42a3-8ec8-fefe61d1e826 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.649680] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8e64b614-31e3-4830-b620-ff8409da8c37 could not be found. [ 633.649826] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 633.649996] env[62600]: INFO nova.compute.manager [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Took 0.04 seconds to destroy the instance on the hypervisor. [ 633.650270] env[62600]: DEBUG oslo.service.loopingcall [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 633.653401] env[62600]: DEBUG nova.compute.manager [-] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 633.653501] env[62600]: DEBUG nova.network.neutron [-] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 633.671484] env[62600]: DEBUG nova.network.neutron [-] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.891475] env[62600]: DEBUG nova.network.neutron [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Successfully created port: b12d3fe3-8bd8-4b35-b17e-b5cc0cf95c9a {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 634.010383] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a12c6f7-1db4-4bff-bee7-63345cb074bd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.018329] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e8eb247-35a8-4898-ab01-89d128e78a66 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.048173] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.048946] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dff3040-b11a-43a1-9126-be8b7e296734 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.056204] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6721b663-fe20-4dff-97a2-f1f7df10be9c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.061593] env[62600]: DEBUG nova.compute.manager [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 634.074527] env[62600]: DEBUG nova.compute.provider_tree [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.142686] env[62600]: DEBUG nova.network.neutron [req-ebe8a09c-4d86-424d-a131-e6d62735614b req-87ddce1f-fdd2-4502-aa03-af8171184845 service nova] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.173604] env[62600]: DEBUG nova.network.neutron [-] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.258811] env[62600]: DEBUG nova.network.neutron [req-ebe8a09c-4d86-424d-a131-e6d62735614b req-87ddce1f-fdd2-4502-aa03-af8171184845 service nova] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.579417] env[62600]: DEBUG nova.scheduler.client.report [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 634.676409] env[62600]: INFO nova.compute.manager [-] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Took 1.02 seconds to deallocate network for instance. [ 634.678939] env[62600]: DEBUG nova.compute.claims [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 634.679162] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.761536] env[62600]: DEBUG oslo_concurrency.lockutils [req-ebe8a09c-4d86-424d-a131-e6d62735614b req-87ddce1f-fdd2-4502-aa03-af8171184845 service nova] Releasing lock "refresh_cache-8e64b614-31e3-4830-b620-ff8409da8c37" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.761820] env[62600]: DEBUG nova.compute.manager [req-ebe8a09c-4d86-424d-a131-e6d62735614b req-87ddce1f-fdd2-4502-aa03-af8171184845 service nova] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Received event network-vif-deleted-cb7cc583-07cb-4cfd-8f10-8d81c32604f8 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 634.792243] env[62600]: ERROR nova.compute.manager [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b12d3fe3-8bd8-4b35-b17e-b5cc0cf95c9a, please check neutron logs for more information. [ 634.792243] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 634.792243] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.792243] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 634.792243] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 634.792243] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 634.792243] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 634.792243] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 634.792243] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.792243] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 634.792243] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.792243] env[62600]: ERROR nova.compute.manager raise self.value [ 634.792243] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 634.792243] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 634.792243] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.792243] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 634.792671] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.792671] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 634.792671] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b12d3fe3-8bd8-4b35-b17e-b5cc0cf95c9a, please check neutron logs for more information. [ 634.792671] env[62600]: ERROR nova.compute.manager [ 634.792671] env[62600]: Traceback (most recent call last): [ 634.792671] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 634.792671] env[62600]: listener.cb(fileno) [ 634.792671] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.792671] env[62600]: result = function(*args, **kwargs) [ 634.792671] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 634.792671] env[62600]: return func(*args, **kwargs) [ 634.792671] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.792671] env[62600]: raise e [ 634.792671] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.792671] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 634.792671] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 634.792671] env[62600]: created_port_ids = self._update_ports_for_instance( [ 634.792671] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 634.792671] env[62600]: with excutils.save_and_reraise_exception(): [ 634.792671] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.792671] env[62600]: self.force_reraise() [ 634.792671] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.792671] env[62600]: raise self.value [ 634.792671] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 634.792671] env[62600]: updated_port = self._update_port( [ 634.792671] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.792671] env[62600]: _ensure_no_port_binding_failure(port) [ 634.792671] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.792671] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 634.793309] env[62600]: nova.exception.PortBindingFailed: Binding failed for port b12d3fe3-8bd8-4b35-b17e-b5cc0cf95c9a, please check neutron logs for more information. [ 634.793309] env[62600]: Removing descriptor: 16 [ 635.073697] env[62600]: DEBUG nova.compute.manager [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 635.084601] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.036s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.085361] env[62600]: ERROR nova.compute.manager [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 660d01e6-2e9a-4002-8ee8-c9b96f07cec5, please check neutron logs for more information. [ 635.085361] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Traceback (most recent call last): [ 635.085361] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 635.085361] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] self.driver.spawn(context, instance, image_meta, [ 635.085361] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 635.085361] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.085361] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.085361] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] vm_ref = self.build_virtual_machine(instance, [ 635.085361] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.085361] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] vif_infos = vmwarevif.get_vif_info(self._session, [ 635.085361] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.085677] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] for vif in network_info: [ 635.085677] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 635.085677] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] return self._sync_wrapper(fn, *args, **kwargs) [ 635.085677] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 635.085677] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] self.wait() [ 635.085677] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 635.085677] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] self[:] = self._gt.wait() [ 635.085677] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.085677] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] return self._exit_event.wait() [ 635.085677] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 635.085677] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] result = hub.switch() [ 635.085677] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 635.085677] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] return self.greenlet.switch() [ 635.085943] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.085943] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] result = function(*args, **kwargs) [ 635.085943] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.085943] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] return func(*args, **kwargs) [ 635.085943] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.085943] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] raise e [ 635.085943] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.085943] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] nwinfo = self.network_api.allocate_for_instance( [ 635.085943] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.085943] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] created_port_ids = self._update_ports_for_instance( [ 635.085943] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.085943] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] with excutils.save_and_reraise_exception(): [ 635.085943] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.086233] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] self.force_reraise() [ 635.086233] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.086233] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] raise self.value [ 635.086233] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.086233] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] updated_port = self._update_port( [ 635.086233] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.086233] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] _ensure_no_port_binding_failure(port) [ 635.086233] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.086233] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] raise exception.PortBindingFailed(port_id=port['id']) [ 635.086233] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] nova.exception.PortBindingFailed: Binding failed for port 660d01e6-2e9a-4002-8ee8-c9b96f07cec5, please check neutron logs for more information. [ 635.086233] env[62600]: ERROR nova.compute.manager [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] [ 635.086489] env[62600]: DEBUG nova.compute.utils [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Binding failed for port 660d01e6-2e9a-4002-8ee8-c9b96f07cec5, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 635.087322] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.156s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.090295] env[62600]: DEBUG nova.compute.manager [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Build of instance ec99a26e-ee0a-44da-80b6-afd1205c7ea0 was re-scheduled: Binding failed for port 660d01e6-2e9a-4002-8ee8-c9b96f07cec5, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 635.090743] env[62600]: DEBUG nova.compute.manager [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 635.090943] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Acquiring lock "refresh_cache-ec99a26e-ee0a-44da-80b6-afd1205c7ea0" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.091173] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Acquired lock "refresh_cache-ec99a26e-ee0a-44da-80b6-afd1205c7ea0" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.091256] env[62600]: DEBUG nova.network.neutron [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 635.099947] env[62600]: DEBUG nova.virt.hardware [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 635.100182] env[62600]: DEBUG nova.virt.hardware [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 635.100337] env[62600]: DEBUG nova.virt.hardware [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 635.100516] env[62600]: DEBUG nova.virt.hardware [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 635.100663] env[62600]: DEBUG nova.virt.hardware [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 635.100811] env[62600]: DEBUG nova.virt.hardware [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 635.101080] env[62600]: DEBUG nova.virt.hardware [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 635.101312] env[62600]: DEBUG nova.virt.hardware [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 635.101505] env[62600]: DEBUG nova.virt.hardware [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 635.101672] env[62600]: DEBUG nova.virt.hardware [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 635.101841] env[62600]: DEBUG nova.virt.hardware [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 635.102745] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ee4fc38-da72-49a7-8c0b-3580fcef34bf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.112105] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672d5929-f889-4d80-9736-09f89fab817c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.127374] env[62600]: ERROR nova.compute.manager [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b12d3fe3-8bd8-4b35-b17e-b5cc0cf95c9a, please check neutron logs for more information. [ 635.127374] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Traceback (most recent call last): [ 635.127374] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 635.127374] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] yield resources [ 635.127374] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 635.127374] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] self.driver.spawn(context, instance, image_meta, [ 635.127374] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 635.127374] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.127374] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.127374] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] vm_ref = self.build_virtual_machine(instance, [ 635.127374] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.127789] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] vif_infos = vmwarevif.get_vif_info(self._session, [ 635.127789] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.127789] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] for vif in network_info: [ 635.127789] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 635.127789] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] return self._sync_wrapper(fn, *args, **kwargs) [ 635.127789] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 635.127789] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] self.wait() [ 635.127789] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 635.127789] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] self[:] = self._gt.wait() [ 635.127789] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.127789] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] return self._exit_event.wait() [ 635.127789] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 635.127789] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] current.throw(*self._exc) [ 635.128226] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.128226] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] result = function(*args, **kwargs) [ 635.128226] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.128226] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] return func(*args, **kwargs) [ 635.128226] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.128226] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] raise e [ 635.128226] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.128226] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] nwinfo = self.network_api.allocate_for_instance( [ 635.128226] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.128226] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] created_port_ids = self._update_ports_for_instance( [ 635.128226] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.128226] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] with excutils.save_and_reraise_exception(): [ 635.128226] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.128650] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] self.force_reraise() [ 635.128650] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.128650] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] raise self.value [ 635.128650] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.128650] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] updated_port = self._update_port( [ 635.128650] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.128650] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] _ensure_no_port_binding_failure(port) [ 635.128650] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.128650] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] raise exception.PortBindingFailed(port_id=port['id']) [ 635.128650] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] nova.exception.PortBindingFailed: Binding failed for port b12d3fe3-8bd8-4b35-b17e-b5cc0cf95c9a, please check neutron logs for more information. [ 635.128650] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] [ 635.128650] env[62600]: INFO nova.compute.manager [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Terminating instance [ 635.129941] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquiring lock "refresh_cache-6467456c-db39-4fd6-b67e-a5be2b803bd3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.130160] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquired lock "refresh_cache-6467456c-db39-4fd6-b67e-a5be2b803bd3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.130334] env[62600]: DEBUG nova.network.neutron [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 635.196859] env[62600]: DEBUG nova.compute.manager [req-0ad04b8f-806a-4264-ad5a-2b04ae94bbb1 req-def36cc6-dbf1-4531-9df8-05399c118212 service nova] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Received event network-changed-b12d3fe3-8bd8-4b35-b17e-b5cc0cf95c9a {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 635.197058] env[62600]: DEBUG nova.compute.manager [req-0ad04b8f-806a-4264-ad5a-2b04ae94bbb1 req-def36cc6-dbf1-4531-9df8-05399c118212 service nova] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Refreshing instance network info cache due to event network-changed-b12d3fe3-8bd8-4b35-b17e-b5cc0cf95c9a. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 635.197245] env[62600]: DEBUG oslo_concurrency.lockutils [req-0ad04b8f-806a-4264-ad5a-2b04ae94bbb1 req-def36cc6-dbf1-4531-9df8-05399c118212 service nova] Acquiring lock "refresh_cache-6467456c-db39-4fd6-b67e-a5be2b803bd3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.612357] env[62600]: DEBUG nova.network.neutron [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.651432] env[62600]: DEBUG nova.network.neutron [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.733094] env[62600]: DEBUG nova.network.neutron [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.765088] env[62600]: DEBUG nova.network.neutron [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.984668] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b224bb9e-121d-4ad5-bbde-2d2f67849976 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.991986] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-604998f3-2a79-4207-9e7c-8974d21c3246 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.022536] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3943fe2e-aaa9-4f4a-8c44-dcf172e32106 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.030347] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2478866f-e664-4d7c-bbe9-684086df1ce7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.043252] env[62600]: DEBUG nova.compute.provider_tree [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.235152] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Releasing lock "refresh_cache-ec99a26e-ee0a-44da-80b6-afd1205c7ea0" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.235398] env[62600]: DEBUG nova.compute.manager [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 636.235601] env[62600]: DEBUG nova.compute.manager [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 636.235772] env[62600]: DEBUG nova.network.neutron [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 636.250745] env[62600]: DEBUG nova.network.neutron [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.269683] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Releasing lock "refresh_cache-6467456c-db39-4fd6-b67e-a5be2b803bd3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.270082] env[62600]: DEBUG nova.compute.manager [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 636.270275] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 636.270594] env[62600]: DEBUG oslo_concurrency.lockutils [req-0ad04b8f-806a-4264-ad5a-2b04ae94bbb1 req-def36cc6-dbf1-4531-9df8-05399c118212 service nova] Acquired lock "refresh_cache-6467456c-db39-4fd6-b67e-a5be2b803bd3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.270778] env[62600]: DEBUG nova.network.neutron [req-0ad04b8f-806a-4264-ad5a-2b04ae94bbb1 req-def36cc6-dbf1-4531-9df8-05399c118212 service nova] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Refreshing network info cache for port b12d3fe3-8bd8-4b35-b17e-b5cc0cf95c9a {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 636.271692] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-12cb1fde-1b11-4aa3-9d33-35367f008225 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.280661] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff03eae5-3dad-4870-b2b6-501c73dd2e6f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.301798] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6467456c-db39-4fd6-b67e-a5be2b803bd3 could not be found. [ 636.302012] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 636.302194] env[62600]: INFO nova.compute.manager [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Took 0.03 seconds to destroy the instance on the hypervisor. [ 636.302427] env[62600]: DEBUG oslo.service.loopingcall [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 636.302647] env[62600]: DEBUG nova.compute.manager [-] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 636.302753] env[62600]: DEBUG nova.network.neutron [-] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 636.318038] env[62600]: DEBUG nova.network.neutron [-] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.545920] env[62600]: DEBUG nova.scheduler.client.report [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 636.753532] env[62600]: DEBUG nova.network.neutron [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.789476] env[62600]: DEBUG nova.network.neutron [req-0ad04b8f-806a-4264-ad5a-2b04ae94bbb1 req-def36cc6-dbf1-4531-9df8-05399c118212 service nova] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.821072] env[62600]: DEBUG nova.network.neutron [-] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.858988] env[62600]: DEBUG nova.network.neutron [req-0ad04b8f-806a-4264-ad5a-2b04ae94bbb1 req-def36cc6-dbf1-4531-9df8-05399c118212 service nova] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.051444] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.964s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.052116] env[62600]: ERROR nova.compute.manager [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 81ec8210-343f-40b7-b92e-ad5a81d2b07e, please check neutron logs for more information. [ 637.052116] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Traceback (most recent call last): [ 637.052116] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 637.052116] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] self.driver.spawn(context, instance, image_meta, [ 637.052116] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 637.052116] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 637.052116] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 637.052116] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] vm_ref = self.build_virtual_machine(instance, [ 637.052116] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 637.052116] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] vif_infos = vmwarevif.get_vif_info(self._session, [ 637.052116] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 637.052398] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] for vif in network_info: [ 637.052398] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 637.052398] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] return self._sync_wrapper(fn, *args, **kwargs) [ 637.052398] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 637.052398] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] self.wait() [ 637.052398] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 637.052398] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] self[:] = self._gt.wait() [ 637.052398] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 637.052398] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] return self._exit_event.wait() [ 637.052398] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 637.052398] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] result = hub.switch() [ 637.052398] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 637.052398] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] return self.greenlet.switch() [ 637.052725] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.052725] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] result = function(*args, **kwargs) [ 637.052725] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 637.052725] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] return func(*args, **kwargs) [ 637.052725] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.052725] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] raise e [ 637.052725] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.052725] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] nwinfo = self.network_api.allocate_for_instance( [ 637.052725] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 637.052725] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] created_port_ids = self._update_ports_for_instance( [ 637.052725] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 637.052725] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] with excutils.save_and_reraise_exception(): [ 637.052725] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.053056] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] self.force_reraise() [ 637.053056] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.053056] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] raise self.value [ 637.053056] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 637.053056] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] updated_port = self._update_port( [ 637.053056] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.053056] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] _ensure_no_port_binding_failure(port) [ 637.053056] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.053056] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] raise exception.PortBindingFailed(port_id=port['id']) [ 637.053056] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] nova.exception.PortBindingFailed: Binding failed for port 81ec8210-343f-40b7-b92e-ad5a81d2b07e, please check neutron logs for more information. [ 637.053056] env[62600]: ERROR nova.compute.manager [instance: a934e86f-1a51-468e-915f-add473fc2bf6] [ 637.053316] env[62600]: DEBUG nova.compute.utils [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Binding failed for port 81ec8210-343f-40b7-b92e-ad5a81d2b07e, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 637.054127] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.490s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.056134] env[62600]: INFO nova.compute.claims [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 637.060867] env[62600]: DEBUG nova.compute.manager [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Build of instance a934e86f-1a51-468e-915f-add473fc2bf6 was re-scheduled: Binding failed for port 81ec8210-343f-40b7-b92e-ad5a81d2b07e, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 637.060867] env[62600]: DEBUG nova.compute.manager [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 637.060867] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Acquiring lock "refresh_cache-a934e86f-1a51-468e-915f-add473fc2bf6" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.060867] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Acquired lock "refresh_cache-a934e86f-1a51-468e-915f-add473fc2bf6" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.061107] env[62600]: DEBUG nova.network.neutron [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 637.261339] env[62600]: INFO nova.compute.manager [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] [instance: ec99a26e-ee0a-44da-80b6-afd1205c7ea0] Took 1.03 seconds to deallocate network for instance. [ 637.323534] env[62600]: INFO nova.compute.manager [-] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Took 1.02 seconds to deallocate network for instance. [ 637.325863] env[62600]: DEBUG nova.compute.claims [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 637.326044] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.361276] env[62600]: DEBUG oslo_concurrency.lockutils [req-0ad04b8f-806a-4264-ad5a-2b04ae94bbb1 req-def36cc6-dbf1-4531-9df8-05399c118212 service nova] Releasing lock "refresh_cache-6467456c-db39-4fd6-b67e-a5be2b803bd3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.361525] env[62600]: DEBUG nova.compute.manager [req-0ad04b8f-806a-4264-ad5a-2b04ae94bbb1 req-def36cc6-dbf1-4531-9df8-05399c118212 service nova] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Received event network-vif-deleted-b12d3fe3-8bd8-4b35-b17e-b5cc0cf95c9a {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 637.582167] env[62600]: DEBUG nova.network.neutron [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.652146] env[62600]: DEBUG nova.network.neutron [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.157524] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Releasing lock "refresh_cache-a934e86f-1a51-468e-915f-add473fc2bf6" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.157524] env[62600]: DEBUG nova.compute.manager [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 638.157524] env[62600]: DEBUG nova.compute.manager [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 638.157524] env[62600]: DEBUG nova.network.neutron [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 638.172537] env[62600]: DEBUG nova.network.neutron [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.288950] env[62600]: INFO nova.scheduler.client.report [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Deleted allocations for instance ec99a26e-ee0a-44da-80b6-afd1205c7ea0 [ 638.453041] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d88bc8d1-1efa-40eb-b9a2-35c395dfa8af {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.460815] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-009b94b6-3ab6-43e3-baba-df5d0bd22c28 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.490422] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-289ce8db-c92e-482f-8cae-900e04a8f3f0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.497564] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a15553e-9f4e-4e07-8ab1-5ec95c2ae2e9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.510326] env[62600]: DEBUG nova.compute.provider_tree [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 638.675602] env[62600]: DEBUG nova.network.neutron [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.798315] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3a6cdf2e-912f-42fc-8227-77b243278597 tempest-ImagesNegativeTestJSON-749837137 tempest-ImagesNegativeTestJSON-749837137-project-member] Lock "ec99a26e-ee0a-44da-80b6-afd1205c7ea0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.796s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.013688] env[62600]: DEBUG nova.scheduler.client.report [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 639.179840] env[62600]: INFO nova.compute.manager [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: a934e86f-1a51-468e-915f-add473fc2bf6] Took 1.02 seconds to deallocate network for instance. [ 639.303227] env[62600]: DEBUG nova.compute.manager [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 639.518104] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.464s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.518646] env[62600]: DEBUG nova.compute.manager [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 639.521381] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.016s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.827210] env[62600]: DEBUG oslo_concurrency.lockutils [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.028193] env[62600]: DEBUG nova.compute.utils [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 640.037053] env[62600]: DEBUG nova.compute.manager [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 640.037162] env[62600]: DEBUG nova.network.neutron [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 640.112960] env[62600]: DEBUG nova.policy [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c4cbb3468ca24fd3b7d65917df95744a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '709246d7fb824ac2951e1c05fea0763b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 640.210206] env[62600]: INFO nova.scheduler.client.report [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Deleted allocations for instance a934e86f-1a51-468e-915f-add473fc2bf6 [ 640.522386] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-245f73b7-6030-48bb-942e-4b76844b2425 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.532868] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e563ef79-eb1c-4506-9fd7-059f41fb9009 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.537175] env[62600]: DEBUG nova.compute.manager [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 640.574252] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2920488c-e540-4cf1-a533-cf11630d9af4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.582888] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3213324e-83cb-41ca-9427-8bc678f2cfff {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.598652] env[62600]: DEBUG nova.compute.provider_tree [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 640.635112] env[62600]: DEBUG nova.network.neutron [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Successfully created port: 0f2d2491-e382-415e-b857-525132aaea97 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 640.723446] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca6b4a3-cbfd-452b-8d15-97b3746f0635 tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Lock "a934e86f-1a51-468e-915f-add473fc2bf6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.852s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.100379] env[62600]: DEBUG nova.scheduler.client.report [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 641.226017] env[62600]: DEBUG nova.compute.manager [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 641.342583] env[62600]: DEBUG nova.compute.manager [req-c47e97eb-ed7f-418c-a499-e70e5c1e7e4c req-62139c12-189d-4058-8d98-0e2546345277 service nova] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Received event network-changed-0f2d2491-e382-415e-b857-525132aaea97 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 641.342788] env[62600]: DEBUG nova.compute.manager [req-c47e97eb-ed7f-418c-a499-e70e5c1e7e4c req-62139c12-189d-4058-8d98-0e2546345277 service nova] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Refreshing instance network info cache due to event network-changed-0f2d2491-e382-415e-b857-525132aaea97. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 641.343175] env[62600]: DEBUG oslo_concurrency.lockutils [req-c47e97eb-ed7f-418c-a499-e70e5c1e7e4c req-62139c12-189d-4058-8d98-0e2546345277 service nova] Acquiring lock "refresh_cache-2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.343175] env[62600]: DEBUG oslo_concurrency.lockutils [req-c47e97eb-ed7f-418c-a499-e70e5c1e7e4c req-62139c12-189d-4058-8d98-0e2546345277 service nova] Acquired lock "refresh_cache-2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.343325] env[62600]: DEBUG nova.network.neutron [req-c47e97eb-ed7f-418c-a499-e70e5c1e7e4c req-62139c12-189d-4058-8d98-0e2546345277 service nova] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Refreshing network info cache for port 0f2d2491-e382-415e-b857-525132aaea97 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 641.525115] env[62600]: ERROR nova.compute.manager [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0f2d2491-e382-415e-b857-525132aaea97, please check neutron logs for more information. [ 641.525115] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 641.525115] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.525115] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 641.525115] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 641.525115] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 641.525115] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 641.525115] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 641.525115] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.525115] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 641.525115] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.525115] env[62600]: ERROR nova.compute.manager raise self.value [ 641.525115] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 641.525115] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 641.525115] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.525115] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 641.525680] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.525680] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 641.525680] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0f2d2491-e382-415e-b857-525132aaea97, please check neutron logs for more information. [ 641.525680] env[62600]: ERROR nova.compute.manager [ 641.525680] env[62600]: Traceback (most recent call last): [ 641.525680] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 641.525680] env[62600]: listener.cb(fileno) [ 641.525680] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.525680] env[62600]: result = function(*args, **kwargs) [ 641.525680] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 641.525680] env[62600]: return func(*args, **kwargs) [ 641.525680] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.525680] env[62600]: raise e [ 641.525680] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.525680] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 641.525680] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 641.525680] env[62600]: created_port_ids = self._update_ports_for_instance( [ 641.525680] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 641.525680] env[62600]: with excutils.save_and_reraise_exception(): [ 641.525680] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.525680] env[62600]: self.force_reraise() [ 641.525680] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.525680] env[62600]: raise self.value [ 641.525680] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 641.525680] env[62600]: updated_port = self._update_port( [ 641.525680] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.525680] env[62600]: _ensure_no_port_binding_failure(port) [ 641.525680] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.525680] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 641.526447] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 0f2d2491-e382-415e-b857-525132aaea97, please check neutron logs for more information. [ 641.526447] env[62600]: Removing descriptor: 15 [ 641.552931] env[62600]: DEBUG nova.compute.manager [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 641.581297] env[62600]: DEBUG nova.virt.hardware [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T12:00:02Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='880954569',id=22,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1815416733',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 641.581559] env[62600]: DEBUG nova.virt.hardware [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 641.581839] env[62600]: DEBUG nova.virt.hardware [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 641.581895] env[62600]: DEBUG nova.virt.hardware [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 641.582035] env[62600]: DEBUG nova.virt.hardware [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 641.582192] env[62600]: DEBUG nova.virt.hardware [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 641.582414] env[62600]: DEBUG nova.virt.hardware [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 641.582637] env[62600]: DEBUG nova.virt.hardware [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 641.582782] env[62600]: DEBUG nova.virt.hardware [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 641.583330] env[62600]: DEBUG nova.virt.hardware [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 641.583330] env[62600]: DEBUG nova.virt.hardware [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 641.584094] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-769eb828-dd67-490d-9582-87b0c6f1034a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.593222] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-669fdf0a-2cfd-4bf6-a937-9c6d6944f0b6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.608637] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.087s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.609243] env[62600]: ERROR nova.compute.manager [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3859d35c-4111-42ef-9ac2-15f4f99f0f49, please check neutron logs for more information. [ 641.609243] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Traceback (most recent call last): [ 641.609243] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 641.609243] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] self.driver.spawn(context, instance, image_meta, [ 641.609243] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 641.609243] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.609243] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.609243] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] vm_ref = self.build_virtual_machine(instance, [ 641.609243] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.609243] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.609243] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.609526] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] for vif in network_info: [ 641.609526] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 641.609526] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] return self._sync_wrapper(fn, *args, **kwargs) [ 641.609526] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 641.609526] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] self.wait() [ 641.609526] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 641.609526] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] self[:] = self._gt.wait() [ 641.609526] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.609526] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] return self._exit_event.wait() [ 641.609526] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 641.609526] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] current.throw(*self._exc) [ 641.609526] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.609526] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] result = function(*args, **kwargs) [ 641.609833] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 641.609833] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] return func(*args, **kwargs) [ 641.609833] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.609833] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] raise e [ 641.609833] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.609833] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] nwinfo = self.network_api.allocate_for_instance( [ 641.609833] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 641.609833] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] created_port_ids = self._update_ports_for_instance( [ 641.609833] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 641.609833] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] with excutils.save_and_reraise_exception(): [ 641.609833] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.609833] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] self.force_reraise() [ 641.609833] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.610138] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] raise self.value [ 641.610138] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 641.610138] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] updated_port = self._update_port( [ 641.610138] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.610138] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] _ensure_no_port_binding_failure(port) [ 641.610138] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.610138] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] raise exception.PortBindingFailed(port_id=port['id']) [ 641.610138] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] nova.exception.PortBindingFailed: Binding failed for port 3859d35c-4111-42ef-9ac2-15f4f99f0f49, please check neutron logs for more information. [ 641.610138] env[62600]: ERROR nova.compute.manager [instance: b3cc82aa-be39-4e89-8972-b5596629961c] [ 641.610138] env[62600]: DEBUG nova.compute.utils [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Binding failed for port 3859d35c-4111-42ef-9ac2-15f4f99f0f49, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 641.611850] env[62600]: ERROR nova.compute.manager [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0f2d2491-e382-415e-b857-525132aaea97, please check neutron logs for more information. [ 641.611850] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Traceback (most recent call last): [ 641.611850] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 641.611850] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] yield resources [ 641.611850] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 641.611850] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] self.driver.spawn(context, instance, image_meta, [ 641.611850] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 641.611850] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 641.611850] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 641.611850] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] vm_ref = self.build_virtual_machine(instance, [ 641.611850] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 641.612153] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] vif_infos = vmwarevif.get_vif_info(self._session, [ 641.612153] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 641.612153] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] for vif in network_info: [ 641.612153] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 641.612153] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] return self._sync_wrapper(fn, *args, **kwargs) [ 641.612153] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 641.612153] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] self.wait() [ 641.612153] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 641.612153] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] self[:] = self._gt.wait() [ 641.612153] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 641.612153] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] return self._exit_event.wait() [ 641.612153] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 641.612153] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] current.throw(*self._exc) [ 641.612528] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 641.612528] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] result = function(*args, **kwargs) [ 641.612528] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 641.612528] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] return func(*args, **kwargs) [ 641.612528] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 641.612528] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] raise e [ 641.612528] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 641.612528] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] nwinfo = self.network_api.allocate_for_instance( [ 641.612528] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 641.612528] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] created_port_ids = self._update_ports_for_instance( [ 641.612528] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 641.612528] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] with excutils.save_and_reraise_exception(): [ 641.612528] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 641.613157] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] self.force_reraise() [ 641.613157] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 641.613157] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] raise self.value [ 641.613157] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 641.613157] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] updated_port = self._update_port( [ 641.613157] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 641.613157] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] _ensure_no_port_binding_failure(port) [ 641.613157] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 641.613157] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] raise exception.PortBindingFailed(port_id=port['id']) [ 641.613157] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] nova.exception.PortBindingFailed: Binding failed for port 0f2d2491-e382-415e-b857-525132aaea97, please check neutron logs for more information. [ 641.613157] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] [ 641.613157] env[62600]: INFO nova.compute.manager [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Terminating instance [ 641.615221] env[62600]: DEBUG nova.compute.manager [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Build of instance b3cc82aa-be39-4e89-8972-b5596629961c was re-scheduled: Binding failed for port 3859d35c-4111-42ef-9ac2-15f4f99f0f49, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 641.615637] env[62600]: DEBUG nova.compute.manager [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 641.615893] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Acquiring lock "refresh_cache-b3cc82aa-be39-4e89-8972-b5596629961c" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.616009] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Acquired lock "refresh_cache-b3cc82aa-be39-4e89-8972-b5596629961c" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.616178] env[62600]: DEBUG nova.network.neutron [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 641.617351] env[62600]: DEBUG oslo_concurrency.lockutils [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.698s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.626836] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Acquiring lock "refresh_cache-2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.758335] env[62600]: DEBUG oslo_concurrency.lockutils [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.861868] env[62600]: DEBUG nova.network.neutron [req-c47e97eb-ed7f-418c-a499-e70e5c1e7e4c req-62139c12-189d-4058-8d98-0e2546345277 service nova] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 641.945407] env[62600]: DEBUG nova.network.neutron [req-c47e97eb-ed7f-418c-a499-e70e5c1e7e4c req-62139c12-189d-4058-8d98-0e2546345277 service nova] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.140161] env[62600]: DEBUG nova.network.neutron [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 642.212284] env[62600]: DEBUG nova.network.neutron [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.447635] env[62600]: DEBUG oslo_concurrency.lockutils [req-c47e97eb-ed7f-418c-a499-e70e5c1e7e4c req-62139c12-189d-4058-8d98-0e2546345277 service nova] Releasing lock "refresh_cache-2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.448036] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Acquired lock "refresh_cache-2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.448254] env[62600]: DEBUG nova.network.neutron [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 642.453175] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94aba216-0041-4dfe-b6c6-0bfcd4814737 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.461398] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67bec341-d391-4ecd-9958-3bdd6594274c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.493493] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-838d6baf-cba0-489c-aa4d-04d382e44ab1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.501016] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e4eb69-4a10-4f52-93e4-d228576fac5b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.514896] env[62600]: DEBUG nova.compute.provider_tree [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.714991] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Releasing lock "refresh_cache-b3cc82aa-be39-4e89-8972-b5596629961c" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.715274] env[62600]: DEBUG nova.compute.manager [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 642.715461] env[62600]: DEBUG nova.compute.manager [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 642.715670] env[62600]: DEBUG nova.network.neutron [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 642.731291] env[62600]: DEBUG nova.network.neutron [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 642.966453] env[62600]: DEBUG nova.network.neutron [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.020025] env[62600]: DEBUG nova.scheduler.client.report [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 643.046744] env[62600]: DEBUG nova.network.neutron [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.233701] env[62600]: DEBUG nova.network.neutron [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.374604] env[62600]: DEBUG nova.compute.manager [req-f8ed3e31-5366-427f-98e8-e1493a063d68 req-2a153f07-db4c-4979-9a62-3301abfe3d3c service nova] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Received event network-vif-deleted-0f2d2491-e382-415e-b857-525132aaea97 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 643.524709] env[62600]: DEBUG oslo_concurrency.lockutils [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.907s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.525394] env[62600]: ERROR nova.compute.manager [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 88b7061a-9a5f-4724-a991-33a74ab51ae7, please check neutron logs for more information. [ 643.525394] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] Traceback (most recent call last): [ 643.525394] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 643.525394] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] self.driver.spawn(context, instance, image_meta, [ 643.525394] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 643.525394] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] self._vmops.spawn(context, instance, image_meta, injected_files, [ 643.525394] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 643.525394] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] vm_ref = self.build_virtual_machine(instance, [ 643.525394] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 643.525394] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] vif_infos = vmwarevif.get_vif_info(self._session, [ 643.525394] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 643.525697] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] for vif in network_info: [ 643.525697] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 643.525697] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] return self._sync_wrapper(fn, *args, **kwargs) [ 643.525697] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 643.525697] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] self.wait() [ 643.525697] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 643.525697] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] self[:] = self._gt.wait() [ 643.525697] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 643.525697] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] return self._exit_event.wait() [ 643.525697] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 643.525697] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] current.throw(*self._exc) [ 643.525697] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 643.525697] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] result = function(*args, **kwargs) [ 643.526039] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 643.526039] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] return func(*args, **kwargs) [ 643.526039] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 643.526039] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] raise e [ 643.526039] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 643.526039] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] nwinfo = self.network_api.allocate_for_instance( [ 643.526039] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 643.526039] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] created_port_ids = self._update_ports_for_instance( [ 643.526039] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 643.526039] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] with excutils.save_and_reraise_exception(): [ 643.526039] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 643.526039] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] self.force_reraise() [ 643.526039] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 643.526371] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] raise self.value [ 643.526371] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 643.526371] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] updated_port = self._update_port( [ 643.526371] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 643.526371] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] _ensure_no_port_binding_failure(port) [ 643.526371] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 643.526371] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] raise exception.PortBindingFailed(port_id=port['id']) [ 643.526371] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] nova.exception.PortBindingFailed: Binding failed for port 88b7061a-9a5f-4724-a991-33a74ab51ae7, please check neutron logs for more information. [ 643.526371] env[62600]: ERROR nova.compute.manager [instance: ba442811-2b0a-44e7-b036-572291aff648] [ 643.526371] env[62600]: DEBUG nova.compute.utils [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Binding failed for port 88b7061a-9a5f-4724-a991-33a74ab51ae7, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 643.527445] env[62600]: DEBUG oslo_concurrency.lockutils [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.036s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.528912] env[62600]: INFO nova.compute.claims [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 643.536960] env[62600]: DEBUG nova.compute.manager [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Build of instance ba442811-2b0a-44e7-b036-572291aff648 was re-scheduled: Binding failed for port 88b7061a-9a5f-4724-a991-33a74ab51ae7, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 643.537435] env[62600]: DEBUG nova.compute.manager [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 643.538703] env[62600]: DEBUG oslo_concurrency.lockutils [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Acquiring lock "refresh_cache-ba442811-2b0a-44e7-b036-572291aff648" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.538703] env[62600]: DEBUG oslo_concurrency.lockutils [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Acquired lock "refresh_cache-ba442811-2b0a-44e7-b036-572291aff648" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.538703] env[62600]: DEBUG nova.network.neutron [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 643.550623] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Releasing lock "refresh_cache-2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.551033] env[62600]: DEBUG nova.compute.manager [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 643.551230] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 643.551525] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d6178d07-1788-4239-ba69-bd99ad239a9c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.561029] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745730e0-7a85-4df8-ab29-80f5b47fd282 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.584131] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a could not be found. [ 643.584368] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 643.584554] env[62600]: INFO nova.compute.manager [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 643.584785] env[62600]: DEBUG oslo.service.loopingcall [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 643.584992] env[62600]: DEBUG nova.compute.manager [-] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.585103] env[62600]: DEBUG nova.network.neutron [-] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 643.600476] env[62600]: DEBUG nova.network.neutron [-] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.736772] env[62600]: INFO nova.compute.manager [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] [instance: b3cc82aa-be39-4e89-8972-b5596629961c] Took 1.02 seconds to deallocate network for instance. [ 644.059025] env[62600]: DEBUG nova.network.neutron [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.104011] env[62600]: DEBUG nova.network.neutron [-] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.145150] env[62600]: DEBUG nova.network.neutron [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.607751] env[62600]: INFO nova.compute.manager [-] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Took 1.02 seconds to deallocate network for instance. [ 644.610143] env[62600]: DEBUG nova.compute.claims [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 644.610324] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.648570] env[62600]: DEBUG oslo_concurrency.lockutils [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Releasing lock "refresh_cache-ba442811-2b0a-44e7-b036-572291aff648" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.648820] env[62600]: DEBUG nova.compute.manager [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 644.649008] env[62600]: DEBUG nova.compute.manager [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 644.649190] env[62600]: DEBUG nova.network.neutron [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 644.665708] env[62600]: DEBUG nova.network.neutron [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.775018] env[62600]: INFO nova.scheduler.client.report [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Deleted allocations for instance b3cc82aa-be39-4e89-8972-b5596629961c [ 644.942176] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff08679-ff2e-4200-85c5-e0dac1950be7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.949658] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c108cd3d-4f03-4eb2-a27a-cd51eeae6b08 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.979269] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60b37e6b-74b4-49b4-8344-660bc17ecaba {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.987287] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a47274a7-393c-4cb2-8b1a-7df71646cac5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.999848] env[62600]: DEBUG nova.compute.provider_tree [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 645.169094] env[62600]: DEBUG nova.network.neutron [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.289233] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c40a25df-1b50-4c09-889a-49805ec6c70c tempest-ServersAdminTestJSON-1319538488 tempest-ServersAdminTestJSON-1319538488-project-member] Lock "b3cc82aa-be39-4e89-8972-b5596629961c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.183s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.502635] env[62600]: DEBUG nova.scheduler.client.report [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 645.670901] env[62600]: INFO nova.compute.manager [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] [instance: ba442811-2b0a-44e7-b036-572291aff648] Took 1.02 seconds to deallocate network for instance. [ 645.792319] env[62600]: DEBUG nova.compute.manager [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 646.007438] env[62600]: DEBUG oslo_concurrency.lockutils [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.007984] env[62600]: DEBUG nova.compute.manager [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 646.010435] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.689s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.314135] env[62600]: DEBUG oslo_concurrency.lockutils [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.515205] env[62600]: DEBUG nova.compute.utils [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 646.523376] env[62600]: DEBUG nova.compute.manager [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 646.523682] env[62600]: DEBUG nova.network.neutron [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 646.574404] env[62600]: DEBUG nova.policy [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4b2ae1429c03426a844cb2bd92d38256', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '583e67fe6a444511946ce9fac5c66420', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 646.726450] env[62600]: INFO nova.scheduler.client.report [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Deleted allocations for instance ba442811-2b0a-44e7-b036-572291aff648 [ 646.892040] env[62600]: DEBUG nova.network.neutron [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Successfully created port: 9cf2f0af-7904-459f-9a8e-a9244e7a5fc7 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 646.960934] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4135aef-bb38-4e7b-b131-9c1f9e9e588e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.970560] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2223150-452c-42a9-aa46-338ba35c179f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.008717] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65ce45eb-0fdb-4039-b0be-8f4f6b502717 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.017774] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c47d30be-5b61-465f-a99b-f80b33ecc41b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.034039] env[62600]: DEBUG nova.compute.manager [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 647.038156] env[62600]: DEBUG nova.compute.provider_tree [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.242636] env[62600]: DEBUG oslo_concurrency.lockutils [None req-855c48ff-85c5-4dc6-965e-9b9f5e1acf7c tempest-AttachInterfacesUnderV243Test-653484817 tempest-AttachInterfacesUnderV243Test-653484817-project-member] Lock "ba442811-2b0a-44e7-b036-572291aff648" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.555s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.545392] env[62600]: DEBUG nova.scheduler.client.report [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 647.745755] env[62600]: DEBUG nova.compute.manager [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 648.057674] env[62600]: DEBUG nova.compute.manager [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 648.058512] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.048s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.059207] env[62600]: ERROR nova.compute.manager [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8871c476-5afb-4342-b4b8-725f0da33941, please check neutron logs for more information. [ 648.059207] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Traceback (most recent call last): [ 648.059207] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 648.059207] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] self.driver.spawn(context, instance, image_meta, [ 648.059207] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 648.059207] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] self._vmops.spawn(context, instance, image_meta, injected_files, [ 648.059207] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 648.059207] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] vm_ref = self.build_virtual_machine(instance, [ 648.059207] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 648.059207] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] vif_infos = vmwarevif.get_vif_info(self._session, [ 648.059207] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 648.059551] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] for vif in network_info: [ 648.059551] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 648.059551] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] return self._sync_wrapper(fn, *args, **kwargs) [ 648.059551] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 648.059551] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] self.wait() [ 648.059551] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 648.059551] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] self[:] = self._gt.wait() [ 648.059551] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 648.059551] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] return self._exit_event.wait() [ 648.059551] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 648.059551] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] current.throw(*self._exc) [ 648.059551] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.059551] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] result = function(*args, **kwargs) [ 648.059904] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.059904] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] return func(*args, **kwargs) [ 648.059904] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.059904] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] raise e [ 648.059904] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.059904] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] nwinfo = self.network_api.allocate_for_instance( [ 648.059904] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.059904] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] created_port_ids = self._update_ports_for_instance( [ 648.059904] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.059904] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] with excutils.save_and_reraise_exception(): [ 648.059904] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.059904] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] self.force_reraise() [ 648.059904] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.060322] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] raise self.value [ 648.060322] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.060322] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] updated_port = self._update_port( [ 648.060322] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.060322] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] _ensure_no_port_binding_failure(port) [ 648.060322] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.060322] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] raise exception.PortBindingFailed(port_id=port['id']) [ 648.060322] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] nova.exception.PortBindingFailed: Binding failed for port 8871c476-5afb-4342-b4b8-725f0da33941, please check neutron logs for more information. [ 648.060322] env[62600]: ERROR nova.compute.manager [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] [ 648.062778] env[62600]: DEBUG nova.compute.utils [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Binding failed for port 8871c476-5afb-4342-b4b8-725f0da33941, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 648.063335] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.932s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.069412] env[62600]: DEBUG nova.compute.manager [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Build of instance b569fb64-27c2-4e20-a2c4-c81058958f98 was re-scheduled: Binding failed for port 8871c476-5afb-4342-b4b8-725f0da33941, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 648.069412] env[62600]: DEBUG nova.compute.manager [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 648.069412] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquiring lock "refresh_cache-b569fb64-27c2-4e20-a2c4-c81058958f98" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.069412] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquired lock "refresh_cache-b569fb64-27c2-4e20-a2c4-c81058958f98" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.070316] env[62600]: DEBUG nova.network.neutron [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 648.097454] env[62600]: DEBUG nova.compute.manager [req-f00453fb-e715-48e1-8f4e-d6e233cca686 req-9aeebadc-6e46-4a8f-ad3b-902323741d3c service nova] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Received event network-changed-9cf2f0af-7904-459f-9a8e-a9244e7a5fc7 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 648.097641] env[62600]: DEBUG nova.compute.manager [req-f00453fb-e715-48e1-8f4e-d6e233cca686 req-9aeebadc-6e46-4a8f-ad3b-902323741d3c service nova] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Refreshing instance network info cache due to event network-changed-9cf2f0af-7904-459f-9a8e-a9244e7a5fc7. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 648.098017] env[62600]: DEBUG oslo_concurrency.lockutils [req-f00453fb-e715-48e1-8f4e-d6e233cca686 req-9aeebadc-6e46-4a8f-ad3b-902323741d3c service nova] Acquiring lock "refresh_cache-222a94b5-7841-490e-9b27-d3f703e6794f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.098017] env[62600]: DEBUG oslo_concurrency.lockutils [req-f00453fb-e715-48e1-8f4e-d6e233cca686 req-9aeebadc-6e46-4a8f-ad3b-902323741d3c service nova] Acquired lock "refresh_cache-222a94b5-7841-490e-9b27-d3f703e6794f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.098348] env[62600]: DEBUG nova.network.neutron [req-f00453fb-e715-48e1-8f4e-d6e233cca686 req-9aeebadc-6e46-4a8f-ad3b-902323741d3c service nova] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Refreshing network info cache for port 9cf2f0af-7904-459f-9a8e-a9244e7a5fc7 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 648.102910] env[62600]: DEBUG nova.virt.hardware [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 648.103193] env[62600]: DEBUG nova.virt.hardware [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 648.103581] env[62600]: DEBUG nova.virt.hardware [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 648.103759] env[62600]: DEBUG nova.virt.hardware [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 648.104011] env[62600]: DEBUG nova.virt.hardware [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 648.104216] env[62600]: DEBUG nova.virt.hardware [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 648.104659] env[62600]: DEBUG nova.virt.hardware [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 648.104878] env[62600]: DEBUG nova.virt.hardware [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 648.105105] env[62600]: DEBUG nova.virt.hardware [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 648.105313] env[62600]: DEBUG nova.virt.hardware [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 648.105601] env[62600]: DEBUG nova.virt.hardware [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 648.106750] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c10e92a-f0f2-4569-af5d-0ac45d500596 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.118631] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f84e046-902c-4803-923a-44113ced4a78 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.136703] env[62600]: ERROR nova.compute.manager [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9cf2f0af-7904-459f-9a8e-a9244e7a5fc7, please check neutron logs for more information. [ 648.136703] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 648.136703] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.136703] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 648.136703] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.136703] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 648.136703] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.136703] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 648.136703] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.136703] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 648.136703] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.136703] env[62600]: ERROR nova.compute.manager raise self.value [ 648.136703] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.136703] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 648.136703] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.136703] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 648.137224] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.137224] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 648.137224] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9cf2f0af-7904-459f-9a8e-a9244e7a5fc7, please check neutron logs for more information. [ 648.137224] env[62600]: ERROR nova.compute.manager [ 648.137224] env[62600]: Traceback (most recent call last): [ 648.137224] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 648.137224] env[62600]: listener.cb(fileno) [ 648.137224] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.137224] env[62600]: result = function(*args, **kwargs) [ 648.137224] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.137224] env[62600]: return func(*args, **kwargs) [ 648.137224] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.137224] env[62600]: raise e [ 648.137224] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.137224] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 648.137224] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.137224] env[62600]: created_port_ids = self._update_ports_for_instance( [ 648.137224] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.137224] env[62600]: with excutils.save_and_reraise_exception(): [ 648.137224] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.137224] env[62600]: self.force_reraise() [ 648.137224] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.137224] env[62600]: raise self.value [ 648.137224] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.137224] env[62600]: updated_port = self._update_port( [ 648.137224] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.137224] env[62600]: _ensure_no_port_binding_failure(port) [ 648.137224] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.137224] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 648.137999] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 9cf2f0af-7904-459f-9a8e-a9244e7a5fc7, please check neutron logs for more information. [ 648.137999] env[62600]: Removing descriptor: 15 [ 648.138170] env[62600]: ERROR nova.compute.manager [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9cf2f0af-7904-459f-9a8e-a9244e7a5fc7, please check neutron logs for more information. [ 648.138170] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Traceback (most recent call last): [ 648.138170] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 648.138170] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] yield resources [ 648.138170] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 648.138170] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] self.driver.spawn(context, instance, image_meta, [ 648.138170] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 648.138170] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 648.138170] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 648.138170] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] vm_ref = self.build_virtual_machine(instance, [ 648.138170] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 648.138471] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] vif_infos = vmwarevif.get_vif_info(self._session, [ 648.138471] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 648.138471] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] for vif in network_info: [ 648.138471] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 648.138471] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] return self._sync_wrapper(fn, *args, **kwargs) [ 648.138471] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 648.138471] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] self.wait() [ 648.138471] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 648.138471] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] self[:] = self._gt.wait() [ 648.138471] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 648.138471] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] return self._exit_event.wait() [ 648.138471] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 648.138471] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] result = hub.switch() [ 648.138786] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 648.138786] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] return self.greenlet.switch() [ 648.138786] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.138786] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] result = function(*args, **kwargs) [ 648.138786] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.138786] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] return func(*args, **kwargs) [ 648.138786] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.138786] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] raise e [ 648.138786] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.138786] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] nwinfo = self.network_api.allocate_for_instance( [ 648.138786] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.138786] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] created_port_ids = self._update_ports_for_instance( [ 648.138786] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.139272] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] with excutils.save_and_reraise_exception(): [ 648.139272] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.139272] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] self.force_reraise() [ 648.139272] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.139272] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] raise self.value [ 648.139272] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.139272] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] updated_port = self._update_port( [ 648.139272] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.139272] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] _ensure_no_port_binding_failure(port) [ 648.139272] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.139272] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] raise exception.PortBindingFailed(port_id=port['id']) [ 648.139272] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] nova.exception.PortBindingFailed: Binding failed for port 9cf2f0af-7904-459f-9a8e-a9244e7a5fc7, please check neutron logs for more information. [ 648.139272] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] [ 648.139627] env[62600]: INFO nova.compute.manager [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Terminating instance [ 648.141189] env[62600]: DEBUG oslo_concurrency.lockutils [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquiring lock "refresh_cache-222a94b5-7841-490e-9b27-d3f703e6794f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.270885] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.381055] env[62600]: DEBUG oslo_concurrency.lockutils [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquiring lock "efff4d72-4673-4a8d-9a81-be9d700ff881" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.381315] env[62600]: DEBUG oslo_concurrency.lockutils [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lock "efff4d72-4673-4a8d-9a81-be9d700ff881" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.050148] env[62600]: DEBUG oslo_concurrency.lockutils [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquiring lock "a358a3cb-deda-419a-aa3c-ce7aeb534240" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.050654] env[62600]: DEBUG oslo_concurrency.lockutils [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lock "a358a3cb-deda-419a-aa3c-ce7aeb534240" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.123220] env[62600]: DEBUG nova.network.neutron [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.127367] env[62600]: DEBUG nova.network.neutron [req-f00453fb-e715-48e1-8f4e-d6e233cca686 req-9aeebadc-6e46-4a8f-ad3b-902323741d3c service nova] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.360029] env[62600]: DEBUG nova.network.neutron [req-f00453fb-e715-48e1-8f4e-d6e233cca686 req-9aeebadc-6e46-4a8f-ad3b-902323741d3c service nova] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.372678] env[62600]: DEBUG nova.network.neutron [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.456483] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe7b832-95ed-473a-be47-d9440926e8fe {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.466539] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d1f941-4a62-45b8-af36-b845eb291188 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.497233] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19fe78ef-6656-4e00-9fdb-e2d79ac646e6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.504660] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52bcc8de-c82a-4fef-92b2-3e6fae241cc7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.522025] env[62600]: DEBUG nova.compute.provider_tree [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.866717] env[62600]: DEBUG oslo_concurrency.lockutils [req-f00453fb-e715-48e1-8f4e-d6e233cca686 req-9aeebadc-6e46-4a8f-ad3b-902323741d3c service nova] Releasing lock "refresh_cache-222a94b5-7841-490e-9b27-d3f703e6794f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.867217] env[62600]: DEBUG oslo_concurrency.lockutils [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquired lock "refresh_cache-222a94b5-7841-490e-9b27-d3f703e6794f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.867406] env[62600]: DEBUG nova.network.neutron [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 649.875054] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Releasing lock "refresh_cache-b569fb64-27c2-4e20-a2c4-c81058958f98" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.875287] env[62600]: DEBUG nova.compute.manager [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 649.875502] env[62600]: DEBUG nova.compute.manager [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 649.875735] env[62600]: DEBUG nova.network.neutron [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 649.895985] env[62600]: DEBUG nova.network.neutron [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 650.024554] env[62600]: DEBUG nova.scheduler.client.report [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 650.134998] env[62600]: DEBUG nova.compute.manager [req-dd4466f9-cc09-4637-8140-e68598996a85 req-337ce490-9426-420a-bfec-88066ea69cb2 service nova] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Received event network-vif-deleted-9cf2f0af-7904-459f-9a8e-a9244e7a5fc7 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 650.392893] env[62600]: DEBUG nova.network.neutron [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 650.398067] env[62600]: DEBUG nova.network.neutron [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.516240] env[62600]: DEBUG nova.network.neutron [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.530149] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.467s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.531065] env[62600]: ERROR nova.compute.manager [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 98868f07-201e-47e6-a446-99bfe9262854, please check neutron logs for more information. [ 650.531065] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Traceback (most recent call last): [ 650.531065] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 650.531065] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] self.driver.spawn(context, instance, image_meta, [ 650.531065] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 650.531065] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] self._vmops.spawn(context, instance, image_meta, injected_files, [ 650.531065] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 650.531065] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] vm_ref = self.build_virtual_machine(instance, [ 650.531065] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 650.531065] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] vif_infos = vmwarevif.get_vif_info(self._session, [ 650.531065] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 650.531955] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] for vif in network_info: [ 650.531955] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 650.531955] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] return self._sync_wrapper(fn, *args, **kwargs) [ 650.531955] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 650.531955] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] self.wait() [ 650.531955] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 650.531955] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] self[:] = self._gt.wait() [ 650.531955] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 650.531955] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] return self._exit_event.wait() [ 650.531955] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 650.531955] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] current.throw(*self._exc) [ 650.531955] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.531955] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] result = function(*args, **kwargs) [ 650.533167] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 650.533167] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] return func(*args, **kwargs) [ 650.533167] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 650.533167] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] raise e [ 650.533167] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.533167] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] nwinfo = self.network_api.allocate_for_instance( [ 650.533167] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 650.533167] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] created_port_ids = self._update_ports_for_instance( [ 650.533167] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 650.533167] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] with excutils.save_and_reraise_exception(): [ 650.533167] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.533167] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] self.force_reraise() [ 650.533167] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.533538] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] raise self.value [ 650.533538] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 650.533538] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] updated_port = self._update_port( [ 650.533538] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.533538] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] _ensure_no_port_binding_failure(port) [ 650.533538] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.533538] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] raise exception.PortBindingFailed(port_id=port['id']) [ 650.533538] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] nova.exception.PortBindingFailed: Binding failed for port 98868f07-201e-47e6-a446-99bfe9262854, please check neutron logs for more information. [ 650.533538] env[62600]: ERROR nova.compute.manager [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] [ 650.533538] env[62600]: DEBUG nova.compute.utils [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Binding failed for port 98868f07-201e-47e6-a446-99bfe9262854, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 650.533878] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.485s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.533878] env[62600]: DEBUG nova.objects.instance [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Lazy-loading 'resources' on Instance uuid 8c0e3a15-0cdd-422a-8a97-95d57621de88 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 650.534343] env[62600]: DEBUG nova.compute.manager [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Build of instance dc77594b-7767-40bf-adbb-c72fa5c80503 was re-scheduled: Binding failed for port 98868f07-201e-47e6-a446-99bfe9262854, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 650.534773] env[62600]: DEBUG nova.compute.manager [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 650.536225] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquiring lock "refresh_cache-dc77594b-7767-40bf-adbb-c72fa5c80503" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.536225] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquired lock "refresh_cache-dc77594b-7767-40bf-adbb-c72fa5c80503" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.536225] env[62600]: DEBUG nova.network.neutron [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 650.900948] env[62600]: INFO nova.compute.manager [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: b569fb64-27c2-4e20-a2c4-c81058958f98] Took 1.03 seconds to deallocate network for instance. [ 651.023234] env[62600]: DEBUG oslo_concurrency.lockutils [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Releasing lock "refresh_cache-222a94b5-7841-490e-9b27-d3f703e6794f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.023406] env[62600]: DEBUG nova.compute.manager [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 651.023524] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 651.023921] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-660205a9-7c40-4b42-8fa2-09c8441b921b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.033161] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-173b9b0d-d69f-46d4-91ea-ec7d144e2921 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.064417] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 222a94b5-7841-490e-9b27-d3f703e6794f could not be found. [ 651.064842] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 651.065204] env[62600]: INFO nova.compute.manager [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 651.066722] env[62600]: DEBUG oslo.service.loopingcall [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 651.067582] env[62600]: DEBUG nova.network.neutron [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 651.070352] env[62600]: DEBUG nova.compute.manager [-] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 651.070451] env[62600]: DEBUG nova.network.neutron [-] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 651.095464] env[62600]: DEBUG nova.network.neutron [-] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 651.171375] env[62600]: DEBUG nova.network.neutron [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.453480] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc1fea08-1835-467f-8ce7-c34ab22130e5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.461757] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80114ef-0b07-4bcf-a85e-3f0abc3b5c31 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.491870] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7341e212-d60c-499e-b605-50f738fa5fd9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.498996] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392ae8bf-0c0f-47ad-8e8e-81f4be914227 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.511818] env[62600]: DEBUG nova.compute.provider_tree [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.598861] env[62600]: DEBUG nova.network.neutron [-] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.675753] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Releasing lock "refresh_cache-dc77594b-7767-40bf-adbb-c72fa5c80503" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.676072] env[62600]: DEBUG nova.compute.manager [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 651.676246] env[62600]: DEBUG nova.compute.manager [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 651.676411] env[62600]: DEBUG nova.network.neutron [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 651.693535] env[62600]: DEBUG nova.network.neutron [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 651.950660] env[62600]: INFO nova.scheduler.client.report [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Deleted allocations for instance b569fb64-27c2-4e20-a2c4-c81058958f98 [ 652.017053] env[62600]: DEBUG nova.scheduler.client.report [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 652.103076] env[62600]: INFO nova.compute.manager [-] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Took 1.03 seconds to deallocate network for instance. [ 652.105101] env[62600]: DEBUG nova.compute.claims [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 652.105286] env[62600]: DEBUG oslo_concurrency.lockutils [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.198406] env[62600]: DEBUG nova.network.neutron [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.460601] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6fe25b0e-f2e4-4f4d-b567-a0f9e9feb8f4 tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Lock "b569fb64-27c2-4e20-a2c4-c81058958f98" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.138s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.522640] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.990s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.524773] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.845s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.540911] env[62600]: INFO nova.scheduler.client.report [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Deleted allocations for instance 8c0e3a15-0cdd-422a-8a97-95d57621de88 [ 652.705062] env[62600]: INFO nova.compute.manager [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: dc77594b-7767-40bf-adbb-c72fa5c80503] Took 1.03 seconds to deallocate network for instance. [ 652.967291] env[62600]: DEBUG nova.compute.manager [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 653.054714] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27acaff5-f6c8-47d2-867b-66c3e6fb22cf tempest-ServerDiagnosticsV248Test-1426798101 tempest-ServerDiagnosticsV248Test-1426798101-project-member] Lock "8c0e3a15-0cdd-422a-8a97-95d57621de88" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.749s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.491552] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.510979] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ffeb7b1-3cb2-4455-a8e6-f6a9d20da7a9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.520216] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc4ce264-6004-4d7e-a680-3d7c4db2735b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.552895] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce61f4e5-c13e-4e6f-96e7-d2f917e9490a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.562291] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d5f6888-f72a-4af3-acc6-33364ef342bd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.577822] env[62600]: DEBUG nova.compute.provider_tree [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.749929] env[62600]: INFO nova.scheduler.client.report [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Deleted allocations for instance dc77594b-7767-40bf-adbb-c72fa5c80503 [ 654.084542] env[62600]: DEBUG nova.scheduler.client.report [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 654.257735] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b1c0b1e5-7fc3-493c-9034-e9960efbb7bc tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Lock "dc77594b-7767-40bf-adbb-c72fa5c80503" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.254s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.590368] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.066s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.592120] env[62600]: ERROR nova.compute.manager [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cb7cc583-07cb-4cfd-8f10-8d81c32604f8, please check neutron logs for more information. [ 654.592120] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Traceback (most recent call last): [ 654.592120] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 654.592120] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] self.driver.spawn(context, instance, image_meta, [ 654.592120] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 654.592120] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] self._vmops.spawn(context, instance, image_meta, injected_files, [ 654.592120] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 654.592120] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] vm_ref = self.build_virtual_machine(instance, [ 654.592120] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 654.592120] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] vif_infos = vmwarevif.get_vif_info(self._session, [ 654.592120] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 654.592456] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] for vif in network_info: [ 654.592456] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 654.592456] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] return self._sync_wrapper(fn, *args, **kwargs) [ 654.592456] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 654.592456] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] self.wait() [ 654.592456] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 654.592456] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] self[:] = self._gt.wait() [ 654.592456] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 654.592456] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] return self._exit_event.wait() [ 654.592456] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 654.592456] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] current.throw(*self._exc) [ 654.592456] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.592456] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] result = function(*args, **kwargs) [ 654.592747] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 654.592747] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] return func(*args, **kwargs) [ 654.592747] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.592747] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] raise e [ 654.592747] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.592747] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] nwinfo = self.network_api.allocate_for_instance( [ 654.592747] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 654.592747] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] created_port_ids = self._update_ports_for_instance( [ 654.592747] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 654.592747] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] with excutils.save_and_reraise_exception(): [ 654.592747] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.592747] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] self.force_reraise() [ 654.592747] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.593043] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] raise self.value [ 654.593043] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 654.593043] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] updated_port = self._update_port( [ 654.593043] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.593043] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] _ensure_no_port_binding_failure(port) [ 654.593043] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.593043] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] raise exception.PortBindingFailed(port_id=port['id']) [ 654.593043] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] nova.exception.PortBindingFailed: Binding failed for port cb7cc583-07cb-4cfd-8f10-8d81c32604f8, please check neutron logs for more information. [ 654.593043] env[62600]: ERROR nova.compute.manager [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] [ 654.593507] env[62600]: DEBUG nova.compute.utils [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Binding failed for port cb7cc583-07cb-4cfd-8f10-8d81c32604f8, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 654.594672] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.269s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.600386] env[62600]: DEBUG nova.compute.manager [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Build of instance 8e64b614-31e3-4830-b620-ff8409da8c37 was re-scheduled: Binding failed for port cb7cc583-07cb-4cfd-8f10-8d81c32604f8, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 654.601532] env[62600]: DEBUG nova.compute.manager [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 654.601781] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Acquiring lock "refresh_cache-8e64b614-31e3-4830-b620-ff8409da8c37" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.601938] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Acquired lock "refresh_cache-8e64b614-31e3-4830-b620-ff8409da8c37" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.602132] env[62600]: DEBUG nova.network.neutron [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 654.763979] env[62600]: DEBUG nova.compute.manager [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 655.138041] env[62600]: DEBUG nova.network.neutron [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 655.249840] env[62600]: DEBUG nova.network.neutron [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.288320] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.520838] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612db14d-eed2-4c84-b6d1-1eb69c6661d6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.530745] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f5511f-3046-4e49-adf5-149524f8ad88 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.559871] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282a8a91-c6af-47f2-9059-df811283debe {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.568896] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e865ce8c-37fa-42d6-ae97-4e01c59b7e37 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.581930] env[62600]: DEBUG nova.compute.provider_tree [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.754925] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Releasing lock "refresh_cache-8e64b614-31e3-4830-b620-ff8409da8c37" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.754925] env[62600]: DEBUG nova.compute.manager [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 655.754925] env[62600]: DEBUG nova.compute.manager [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 655.754925] env[62600]: DEBUG nova.network.neutron [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 655.781782] env[62600]: DEBUG nova.network.neutron [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 656.085494] env[62600]: DEBUG nova.scheduler.client.report [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 656.284976] env[62600]: DEBUG nova.network.neutron [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.346034] env[62600]: DEBUG oslo_concurrency.lockutils [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Acquiring lock "e25c631d-3d1b-40d2-9fb1-a65431f991a9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.346034] env[62600]: DEBUG oslo_concurrency.lockutils [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Lock "e25c631d-3d1b-40d2-9fb1-a65431f991a9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.589856] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.995s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.590667] env[62600]: ERROR nova.compute.manager [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b12d3fe3-8bd8-4b35-b17e-b5cc0cf95c9a, please check neutron logs for more information. [ 656.590667] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Traceback (most recent call last): [ 656.590667] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.590667] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] self.driver.spawn(context, instance, image_meta, [ 656.590667] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 656.590667] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.590667] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.590667] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] vm_ref = self.build_virtual_machine(instance, [ 656.590667] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.590667] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.590667] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.591054] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] for vif in network_info: [ 656.591054] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.591054] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] return self._sync_wrapper(fn, *args, **kwargs) [ 656.591054] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.591054] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] self.wait() [ 656.591054] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.591054] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] self[:] = self._gt.wait() [ 656.591054] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.591054] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] return self._exit_event.wait() [ 656.591054] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 656.591054] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] current.throw(*self._exc) [ 656.591054] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.591054] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] result = function(*args, **kwargs) [ 656.591455] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 656.591455] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] return func(*args, **kwargs) [ 656.591455] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.591455] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] raise e [ 656.591455] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.591455] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] nwinfo = self.network_api.allocate_for_instance( [ 656.591455] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 656.591455] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] created_port_ids = self._update_ports_for_instance( [ 656.591455] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 656.591455] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] with excutils.save_and_reraise_exception(): [ 656.591455] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.591455] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] self.force_reraise() [ 656.591455] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.591797] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] raise self.value [ 656.591797] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 656.591797] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] updated_port = self._update_port( [ 656.591797] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.591797] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] _ensure_no_port_binding_failure(port) [ 656.591797] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.591797] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] raise exception.PortBindingFailed(port_id=port['id']) [ 656.591797] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] nova.exception.PortBindingFailed: Binding failed for port b12d3fe3-8bd8-4b35-b17e-b5cc0cf95c9a, please check neutron logs for more information. [ 656.591797] env[62600]: ERROR nova.compute.manager [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] [ 656.591797] env[62600]: DEBUG nova.compute.utils [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Binding failed for port b12d3fe3-8bd8-4b35-b17e-b5cc0cf95c9a, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 656.592454] env[62600]: DEBUG oslo_concurrency.lockutils [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.765s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.593884] env[62600]: INFO nova.compute.claims [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 656.596593] env[62600]: DEBUG nova.compute.manager [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Build of instance 6467456c-db39-4fd6-b67e-a5be2b803bd3 was re-scheduled: Binding failed for port b12d3fe3-8bd8-4b35-b17e-b5cc0cf95c9a, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 656.597052] env[62600]: DEBUG nova.compute.manager [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 656.597288] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquiring lock "refresh_cache-6467456c-db39-4fd6-b67e-a5be2b803bd3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.597440] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Acquired lock "refresh_cache-6467456c-db39-4fd6-b67e-a5be2b803bd3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.597609] env[62600]: DEBUG nova.network.neutron [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 656.788376] env[62600]: INFO nova.compute.manager [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] [instance: 8e64b614-31e3-4830-b620-ff8409da8c37] Took 1.03 seconds to deallocate network for instance. [ 656.978141] env[62600]: DEBUG oslo_concurrency.lockutils [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Acquiring lock "0a8840f1-7681-4aaa-9f3a-32b72c04c1c4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.978363] env[62600]: DEBUG oslo_concurrency.lockutils [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Lock "0a8840f1-7681-4aaa-9f3a-32b72c04c1c4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.124985] env[62600]: DEBUG nova.network.neutron [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.213441] env[62600]: DEBUG nova.network.neutron [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.647732] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 657.648231] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 657.717384] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Releasing lock "refresh_cache-6467456c-db39-4fd6-b67e-a5be2b803bd3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.717384] env[62600]: DEBUG nova.compute.manager [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 657.717384] env[62600]: DEBUG nova.compute.manager [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 657.717847] env[62600]: DEBUG nova.network.neutron [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 657.735368] env[62600]: DEBUG nova.network.neutron [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.816393] env[62600]: INFO nova.scheduler.client.report [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Deleted allocations for instance 8e64b614-31e3-4830-b620-ff8409da8c37 [ 657.977120] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-501c941b-e9b4-4653-a42d-f831bc830862 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.984484] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5eb94e4-23cb-4c83-ad52-8561d7d52285 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.014378] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-910b430b-7ef8-41dd-84f1-68693356feec {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.021183] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0147e7cf-c791-40ca-ae63-c1faea28da36 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.034732] env[62600]: DEBUG nova.compute.provider_tree [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.154348] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 658.154348] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Starting heal instance info cache {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 658.154348] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Rebuilding the list of instances to heal {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 658.239201] env[62600]: DEBUG nova.network.neutron [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.326267] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7b63ec62-38b6-437a-8dc4-bbe7cc51d4a7 tempest-VolumesAssistedSnapshotsTest-1584003358 tempest-VolumesAssistedSnapshotsTest-1584003358-project-member] Lock "8e64b614-31e3-4830-b620-ff8409da8c37" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.226s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.538076] env[62600]: DEBUG nova.scheduler.client.report [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 658.657336] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Skipping network cache update for instance because it is Building. {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 658.657762] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Skipping network cache update for instance because it is Building. {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 658.657762] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Skipping network cache update for instance because it is Building. {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 658.657762] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Didn't find any instances for network info cache update. {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 658.657933] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 658.658122] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 658.658243] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 658.658481] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 658.658520] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 658.658629] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 658.658755] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62600) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 658.658936] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 658.740805] env[62600]: INFO nova.compute.manager [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] [instance: 6467456c-db39-4fd6-b67e-a5be2b803bd3] Took 1.02 seconds to deallocate network for instance. [ 658.828991] env[62600]: DEBUG nova.compute.manager [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 659.043018] env[62600]: DEBUG oslo_concurrency.lockutils [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.043533] env[62600]: DEBUG nova.compute.manager [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 659.046182] env[62600]: DEBUG oslo_concurrency.lockutils [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.288s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.050238] env[62600]: INFO nova.compute.claims [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 659.161529] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.358494] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.556170] env[62600]: DEBUG nova.compute.utils [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 659.560493] env[62600]: DEBUG nova.compute.manager [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 659.560684] env[62600]: DEBUG nova.network.neutron [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 659.608566] env[62600]: DEBUG nova.policy [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd088efa40c4641159ec90432aabebdb8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ca9c66a916a5480ab1400c3c562fe447', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 659.771800] env[62600]: INFO nova.scheduler.client.report [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Deleted allocations for instance 6467456c-db39-4fd6-b67e-a5be2b803bd3 [ 659.874280] env[62600]: DEBUG nova.network.neutron [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Successfully created port: e4f492e1-f18e-4fbd-8c1f-bab956693c6d {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 660.063509] env[62600]: DEBUG nova.compute.manager [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 660.279872] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9dab582a-97cf-44ad-8264-7eb02c4cee4a tempest-ListServerFiltersTestJSON-840819745 tempest-ListServerFiltersTestJSON-840819745-project-member] Lock "6467456c-db39-4fd6-b67e-a5be2b803bd3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.650s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.478941] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e063bc-ead8-4f6f-b269-5c362e0ab89d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.489456] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a734e12-b6b6-4b6a-a9a9-26011ab1e9d8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.520234] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ed7ce3e-a17c-424a-a1db-6d443072f1e5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.528134] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d89f71-1c8d-4151-a22c-3a65a23cec6c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.541809] env[62600]: DEBUG nova.compute.provider_tree [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.782905] env[62600]: DEBUG nova.compute.manager [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 661.036608] env[62600]: DEBUG nova.compute.manager [req-f6e4e3a8-8ca0-43aa-908c-ec85b3c78f18 req-728558f1-ebdd-47f5-b93a-d6dc01f75c2a service nova] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Received event network-changed-e4f492e1-f18e-4fbd-8c1f-bab956693c6d {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 661.036899] env[62600]: DEBUG nova.compute.manager [req-f6e4e3a8-8ca0-43aa-908c-ec85b3c78f18 req-728558f1-ebdd-47f5-b93a-d6dc01f75c2a service nova] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Refreshing instance network info cache due to event network-changed-e4f492e1-f18e-4fbd-8c1f-bab956693c6d. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 661.037070] env[62600]: DEBUG oslo_concurrency.lockutils [req-f6e4e3a8-8ca0-43aa-908c-ec85b3c78f18 req-728558f1-ebdd-47f5-b93a-d6dc01f75c2a service nova] Acquiring lock "refresh_cache-4455c247-a707-4c3a-9323-0110a1290780" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.037184] env[62600]: DEBUG oslo_concurrency.lockutils [req-f6e4e3a8-8ca0-43aa-908c-ec85b3c78f18 req-728558f1-ebdd-47f5-b93a-d6dc01f75c2a service nova] Acquired lock "refresh_cache-4455c247-a707-4c3a-9323-0110a1290780" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.037340] env[62600]: DEBUG nova.network.neutron [req-f6e4e3a8-8ca0-43aa-908c-ec85b3c78f18 req-728558f1-ebdd-47f5-b93a-d6dc01f75c2a service nova] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Refreshing network info cache for port e4f492e1-f18e-4fbd-8c1f-bab956693c6d {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 661.045677] env[62600]: DEBUG nova.scheduler.client.report [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.072605] env[62600]: DEBUG nova.compute.manager [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 661.105561] env[62600]: DEBUG nova.virt.hardware [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 661.106054] env[62600]: DEBUG nova.virt.hardware [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 661.106238] env[62600]: DEBUG nova.virt.hardware [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 661.106796] env[62600]: DEBUG nova.virt.hardware [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 661.106796] env[62600]: DEBUG nova.virt.hardware [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 661.106796] env[62600]: DEBUG nova.virt.hardware [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 661.106949] env[62600]: DEBUG nova.virt.hardware [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 661.107130] env[62600]: DEBUG nova.virt.hardware [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 661.107449] env[62600]: DEBUG nova.virt.hardware [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 661.107520] env[62600]: DEBUG nova.virt.hardware [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 661.107690] env[62600]: DEBUG nova.virt.hardware [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 661.109414] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a3a4c3-0d47-4ccd-b0a6-513731177ef2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.119100] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39628a17-cee8-4320-9b84-a04f9d3b24bf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.310136] env[62600]: DEBUG oslo_concurrency.lockutils [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.550155] env[62600]: ERROR nova.compute.manager [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e4f492e1-f18e-4fbd-8c1f-bab956693c6d, please check neutron logs for more information. [ 661.550155] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 661.550155] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.550155] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 661.550155] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.550155] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 661.550155] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.550155] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 661.550155] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.550155] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 661.550155] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.550155] env[62600]: ERROR nova.compute.manager raise self.value [ 661.550155] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.550155] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 661.550155] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.550155] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 661.550592] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.550592] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 661.550592] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e4f492e1-f18e-4fbd-8c1f-bab956693c6d, please check neutron logs for more information. [ 661.550592] env[62600]: ERROR nova.compute.manager [ 661.550592] env[62600]: Traceback (most recent call last): [ 661.550592] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 661.550592] env[62600]: listener.cb(fileno) [ 661.550592] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.550592] env[62600]: result = function(*args, **kwargs) [ 661.550592] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 661.550592] env[62600]: return func(*args, **kwargs) [ 661.550592] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.550592] env[62600]: raise e [ 661.550592] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.550592] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 661.550592] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.550592] env[62600]: created_port_ids = self._update_ports_for_instance( [ 661.550592] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.550592] env[62600]: with excutils.save_and_reraise_exception(): [ 661.550592] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.550592] env[62600]: self.force_reraise() [ 661.550592] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.550592] env[62600]: raise self.value [ 661.550592] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.550592] env[62600]: updated_port = self._update_port( [ 661.550592] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.550592] env[62600]: _ensure_no_port_binding_failure(port) [ 661.550592] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.550592] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 661.551366] env[62600]: nova.exception.PortBindingFailed: Binding failed for port e4f492e1-f18e-4fbd-8c1f-bab956693c6d, please check neutron logs for more information. [ 661.551366] env[62600]: Removing descriptor: 15 [ 661.551929] env[62600]: DEBUG oslo_concurrency.lockutils [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.506s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.553117] env[62600]: DEBUG nova.compute.manager [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 661.556822] env[62600]: ERROR nova.compute.manager [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e4f492e1-f18e-4fbd-8c1f-bab956693c6d, please check neutron logs for more information. [ 661.556822] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] Traceback (most recent call last): [ 661.556822] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 661.556822] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] yield resources [ 661.556822] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 661.556822] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] self.driver.spawn(context, instance, image_meta, [ 661.556822] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 661.556822] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] self._vmops.spawn(context, instance, image_meta, injected_files, [ 661.556822] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 661.556822] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] vm_ref = self.build_virtual_machine(instance, [ 661.556822] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 661.557158] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] vif_infos = vmwarevif.get_vif_info(self._session, [ 661.557158] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 661.557158] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] for vif in network_info: [ 661.557158] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 661.557158] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] return self._sync_wrapper(fn, *args, **kwargs) [ 661.557158] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 661.557158] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] self.wait() [ 661.557158] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 661.557158] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] self[:] = self._gt.wait() [ 661.557158] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 661.557158] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] return self._exit_event.wait() [ 661.557158] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 661.557158] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] result = hub.switch() [ 661.557501] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 661.557501] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] return self.greenlet.switch() [ 661.557501] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.557501] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] result = function(*args, **kwargs) [ 661.557501] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 661.557501] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] return func(*args, **kwargs) [ 661.557501] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.557501] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] raise e [ 661.557501] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.557501] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] nwinfo = self.network_api.allocate_for_instance( [ 661.557501] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.557501] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] created_port_ids = self._update_ports_for_instance( [ 661.557501] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.557857] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] with excutils.save_and_reraise_exception(): [ 661.557857] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.557857] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] self.force_reraise() [ 661.557857] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.557857] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] raise self.value [ 661.557857] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.557857] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] updated_port = self._update_port( [ 661.557857] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.557857] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] _ensure_no_port_binding_failure(port) [ 661.557857] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.557857] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] raise exception.PortBindingFailed(port_id=port['id']) [ 661.557857] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] nova.exception.PortBindingFailed: Binding failed for port e4f492e1-f18e-4fbd-8c1f-bab956693c6d, please check neutron logs for more information. [ 661.557857] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] [ 661.558227] env[62600]: INFO nova.compute.manager [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Terminating instance [ 661.560275] env[62600]: DEBUG oslo_concurrency.lockutils [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Acquiring lock "refresh_cache-4455c247-a707-4c3a-9323-0110a1290780" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.561072] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.951s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.574291] env[62600]: DEBUG nova.network.neutron [req-f6e4e3a8-8ca0-43aa-908c-ec85b3c78f18 req-728558f1-ebdd-47f5-b93a-d6dc01f75c2a service nova] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 661.713173] env[62600]: DEBUG nova.network.neutron [req-f6e4e3a8-8ca0-43aa-908c-ec85b3c78f18 req-728558f1-ebdd-47f5-b93a-d6dc01f75c2a service nova] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.063490] env[62600]: DEBUG nova.compute.utils [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 662.068314] env[62600]: DEBUG nova.compute.manager [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 662.068314] env[62600]: DEBUG nova.network.neutron [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 662.172835] env[62600]: DEBUG nova.policy [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ad7c90a5d6744516bb01244ea7bc4854', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cf883f207a8c48fa8dce3a2821dd8610', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 662.214665] env[62600]: DEBUG oslo_concurrency.lockutils [req-f6e4e3a8-8ca0-43aa-908c-ec85b3c78f18 req-728558f1-ebdd-47f5-b93a-d6dc01f75c2a service nova] Releasing lock "refresh_cache-4455c247-a707-4c3a-9323-0110a1290780" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.216070] env[62600]: DEBUG oslo_concurrency.lockutils [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Acquired lock "refresh_cache-4455c247-a707-4c3a-9323-0110a1290780" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.216070] env[62600]: DEBUG nova.network.neutron [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 662.502272] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c9354e9-e6af-4b13-afca-68cee6bf1b0f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.510460] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e6d0b78-c3d0-49bd-8acb-783137cb11a1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.546402] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03846828-fae5-4bac-ac7c-705a643a20c4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.555759] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a2d4e42-d9e5-40a3-8274-649a99ba4e46 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.570661] env[62600]: DEBUG nova.compute.manager [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 662.573601] env[62600]: DEBUG nova.compute.provider_tree [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.676222] env[62600]: DEBUG nova.network.neutron [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Successfully created port: 556c03da-1c62-48dc-a7f4-d107804f7a59 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 662.737235] env[62600]: DEBUG nova.network.neutron [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.906996] env[62600]: DEBUG nova.network.neutron [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.080674] env[62600]: DEBUG nova.scheduler.client.report [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 663.242550] env[62600]: DEBUG nova.compute.manager [req-d2118dcd-731d-4357-903b-1646495686ef req-f24181ea-2867-4fc2-8e2b-501dfb2bd0b9 service nova] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Received event network-vif-deleted-e4f492e1-f18e-4fbd-8c1f-bab956693c6d {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 663.409671] env[62600]: DEBUG oslo_concurrency.lockutils [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Releasing lock "refresh_cache-4455c247-a707-4c3a-9323-0110a1290780" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.410148] env[62600]: DEBUG nova.compute.manager [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 663.410355] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 663.410646] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9533a2df-38d7-41d0-9da0-c958c674dd71 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.419674] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88286e5e-6e04-420e-9de5-1e2e28c7fff3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.441276] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4455c247-a707-4c3a-9323-0110a1290780 could not be found. [ 663.441497] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 663.441677] env[62600]: INFO nova.compute.manager [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Took 0.03 seconds to destroy the instance on the hypervisor. [ 663.441922] env[62600]: DEBUG oslo.service.loopingcall [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 663.442152] env[62600]: DEBUG nova.compute.manager [-] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 663.442246] env[62600]: DEBUG nova.network.neutron [-] [instance: 4455c247-a707-4c3a-9323-0110a1290780] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 663.471416] env[62600]: DEBUG nova.network.neutron [-] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 663.586722] env[62600]: DEBUG nova.compute.manager [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 663.593907] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.030s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.593907] env[62600]: ERROR nova.compute.manager [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0f2d2491-e382-415e-b857-525132aaea97, please check neutron logs for more information. [ 663.593907] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Traceback (most recent call last): [ 663.593907] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 663.593907] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] self.driver.spawn(context, instance, image_meta, [ 663.593907] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 663.593907] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 663.593907] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 663.593907] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] vm_ref = self.build_virtual_machine(instance, [ 663.594431] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 663.594431] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] vif_infos = vmwarevif.get_vif_info(self._session, [ 663.594431] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 663.594431] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] for vif in network_info: [ 663.594431] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 663.594431] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] return self._sync_wrapper(fn, *args, **kwargs) [ 663.594431] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 663.594431] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] self.wait() [ 663.594431] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 663.594431] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] self[:] = self._gt.wait() [ 663.594431] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 663.594431] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] return self._exit_event.wait() [ 663.594431] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 663.594925] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] current.throw(*self._exc) [ 663.594925] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.594925] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] result = function(*args, **kwargs) [ 663.594925] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 663.594925] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] return func(*args, **kwargs) [ 663.594925] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.594925] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] raise e [ 663.594925] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.594925] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] nwinfo = self.network_api.allocate_for_instance( [ 663.594925] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 663.594925] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] created_port_ids = self._update_ports_for_instance( [ 663.594925] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 663.594925] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] with excutils.save_and_reraise_exception(): [ 663.595420] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.595420] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] self.force_reraise() [ 663.595420] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.595420] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] raise self.value [ 663.595420] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 663.595420] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] updated_port = self._update_port( [ 663.595420] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.595420] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] _ensure_no_port_binding_failure(port) [ 663.595420] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.595420] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] raise exception.PortBindingFailed(port_id=port['id']) [ 663.595420] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] nova.exception.PortBindingFailed: Binding failed for port 0f2d2491-e382-415e-b857-525132aaea97, please check neutron logs for more information. [ 663.595420] env[62600]: ERROR nova.compute.manager [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] [ 663.596268] env[62600]: DEBUG nova.compute.utils [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Binding failed for port 0f2d2491-e382-415e-b857-525132aaea97, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 663.596268] env[62600]: DEBUG oslo_concurrency.lockutils [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.281s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.596268] env[62600]: INFO nova.compute.claims [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 663.602067] env[62600]: DEBUG nova.compute.manager [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Build of instance 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a was re-scheduled: Binding failed for port 0f2d2491-e382-415e-b857-525132aaea97, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 663.604606] env[62600]: DEBUG nova.compute.manager [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 663.604876] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Acquiring lock "refresh_cache-2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.605109] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Acquired lock "refresh_cache-2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.605212] env[62600]: DEBUG nova.network.neutron [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 663.630107] env[62600]: DEBUG nova.virt.hardware [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 663.630368] env[62600]: DEBUG nova.virt.hardware [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 663.630529] env[62600]: DEBUG nova.virt.hardware [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 663.630708] env[62600]: DEBUG nova.virt.hardware [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 663.630854] env[62600]: DEBUG nova.virt.hardware [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 663.631033] env[62600]: DEBUG nova.virt.hardware [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 663.631225] env[62600]: DEBUG nova.virt.hardware [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 663.631384] env[62600]: DEBUG nova.virt.hardware [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 663.631567] env[62600]: DEBUG nova.virt.hardware [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 663.631712] env[62600]: DEBUG nova.virt.hardware [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 663.631883] env[62600]: DEBUG nova.virt.hardware [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 663.633172] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9758b4fd-b1df-4149-9ea9-349e34b87f86 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.642911] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c769ee-567e-4ed3-a407-5a5533d2a08b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.973850] env[62600]: DEBUG nova.network.neutron [-] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.142969] env[62600]: DEBUG nova.network.neutron [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 664.393479] env[62600]: DEBUG nova.network.neutron [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.476131] env[62600]: INFO nova.compute.manager [-] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Took 1.03 seconds to deallocate network for instance. [ 664.478662] env[62600]: DEBUG nova.compute.claims [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 664.479194] env[62600]: DEBUG oslo_concurrency.lockutils [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.847442] env[62600]: ERROR nova.compute.manager [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 556c03da-1c62-48dc-a7f4-d107804f7a59, please check neutron logs for more information. [ 664.847442] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 664.847442] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.847442] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 664.847442] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 664.847442] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 664.847442] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 664.847442] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 664.847442] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.847442] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 664.847442] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.847442] env[62600]: ERROR nova.compute.manager raise self.value [ 664.847442] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 664.847442] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 664.847442] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.847442] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 664.847897] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.847897] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 664.847897] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 556c03da-1c62-48dc-a7f4-d107804f7a59, please check neutron logs for more information. [ 664.847897] env[62600]: ERROR nova.compute.manager [ 664.847897] env[62600]: Traceback (most recent call last): [ 664.847897] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 664.847897] env[62600]: listener.cb(fileno) [ 664.847897] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.847897] env[62600]: result = function(*args, **kwargs) [ 664.847897] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 664.847897] env[62600]: return func(*args, **kwargs) [ 664.847897] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.847897] env[62600]: raise e [ 664.847897] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.847897] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 664.847897] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 664.847897] env[62600]: created_port_ids = self._update_ports_for_instance( [ 664.847897] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 664.847897] env[62600]: with excutils.save_and_reraise_exception(): [ 664.847897] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.847897] env[62600]: self.force_reraise() [ 664.847897] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.847897] env[62600]: raise self.value [ 664.847897] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 664.847897] env[62600]: updated_port = self._update_port( [ 664.847897] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.847897] env[62600]: _ensure_no_port_binding_failure(port) [ 664.847897] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.847897] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 664.848659] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 556c03da-1c62-48dc-a7f4-d107804f7a59, please check neutron logs for more information. [ 664.848659] env[62600]: Removing descriptor: 16 [ 664.848659] env[62600]: ERROR nova.compute.manager [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 556c03da-1c62-48dc-a7f4-d107804f7a59, please check neutron logs for more information. [ 664.848659] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Traceback (most recent call last): [ 664.848659] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 664.848659] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] yield resources [ 664.848659] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 664.848659] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] self.driver.spawn(context, instance, image_meta, [ 664.848659] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 664.848659] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 664.848659] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 664.848659] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] vm_ref = self.build_virtual_machine(instance, [ 664.849017] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 664.849017] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] vif_infos = vmwarevif.get_vif_info(self._session, [ 664.849017] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 664.849017] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] for vif in network_info: [ 664.849017] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 664.849017] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] return self._sync_wrapper(fn, *args, **kwargs) [ 664.849017] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 664.849017] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] self.wait() [ 664.849017] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 664.849017] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] self[:] = self._gt.wait() [ 664.849017] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 664.849017] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] return self._exit_event.wait() [ 664.849017] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 664.849409] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] result = hub.switch() [ 664.849409] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 664.849409] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] return self.greenlet.switch() [ 664.849409] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.849409] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] result = function(*args, **kwargs) [ 664.849409] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 664.849409] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] return func(*args, **kwargs) [ 664.849409] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.849409] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] raise e [ 664.849409] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.849409] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] nwinfo = self.network_api.allocate_for_instance( [ 664.849409] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 664.849409] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] created_port_ids = self._update_ports_for_instance( [ 664.849845] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 664.849845] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] with excutils.save_and_reraise_exception(): [ 664.849845] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.849845] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] self.force_reraise() [ 664.849845] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.849845] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] raise self.value [ 664.849845] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 664.849845] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] updated_port = self._update_port( [ 664.849845] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.849845] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] _ensure_no_port_binding_failure(port) [ 664.849845] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.849845] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] raise exception.PortBindingFailed(port_id=port['id']) [ 664.850324] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] nova.exception.PortBindingFailed: Binding failed for port 556c03da-1c62-48dc-a7f4-d107804f7a59, please check neutron logs for more information. [ 664.850324] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] [ 664.850324] env[62600]: INFO nova.compute.manager [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Terminating instance [ 664.850826] env[62600]: DEBUG oslo_concurrency.lockutils [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Acquiring lock "refresh_cache-09ccf23b-4763-47e1-afad-59e2b06b2d9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.850988] env[62600]: DEBUG oslo_concurrency.lockutils [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Acquired lock "refresh_cache-09ccf23b-4763-47e1-afad-59e2b06b2d9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.851168] env[62600]: DEBUG nova.network.neutron [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 664.900776] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Releasing lock "refresh_cache-2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.901022] env[62600]: DEBUG nova.compute.manager [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 664.901203] env[62600]: DEBUG nova.compute.manager [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 664.901361] env[62600]: DEBUG nova.network.neutron [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 664.920480] env[62600]: DEBUG nova.network.neutron [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.078929] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c489f139-23c4-408d-a7e2-aea8e4b3e104 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.087227] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5209042-16a3-48ce-8b80-0fda7c51e41b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.123554] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e156223-c4ec-4c09-93cb-4792fb8fb5e0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.131793] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f11c49-fce5-4c24-b786-0dea0493bd94 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.147545] env[62600]: DEBUG nova.compute.provider_tree [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.295854] env[62600]: DEBUG nova.compute.manager [req-b0750c52-1e37-49f4-bd4c-85d1dcf910d0 req-b1011878-95c5-4fcc-929e-e12ec2ab6092 service nova] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Received event network-changed-556c03da-1c62-48dc-a7f4-d107804f7a59 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 665.296099] env[62600]: DEBUG nova.compute.manager [req-b0750c52-1e37-49f4-bd4c-85d1dcf910d0 req-b1011878-95c5-4fcc-929e-e12ec2ab6092 service nova] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Refreshing instance network info cache due to event network-changed-556c03da-1c62-48dc-a7f4-d107804f7a59. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 665.296324] env[62600]: DEBUG oslo_concurrency.lockutils [req-b0750c52-1e37-49f4-bd4c-85d1dcf910d0 req-b1011878-95c5-4fcc-929e-e12ec2ab6092 service nova] Acquiring lock "refresh_cache-09ccf23b-4763-47e1-afad-59e2b06b2d9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.376842] env[62600]: DEBUG nova.network.neutron [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.424123] env[62600]: DEBUG nova.network.neutron [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.452153] env[62600]: DEBUG nova.network.neutron [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.653019] env[62600]: DEBUG nova.scheduler.client.report [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 665.927073] env[62600]: INFO nova.compute.manager [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] [instance: 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a] Took 1.03 seconds to deallocate network for instance. [ 665.954978] env[62600]: DEBUG oslo_concurrency.lockutils [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Releasing lock "refresh_cache-09ccf23b-4763-47e1-afad-59e2b06b2d9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.955479] env[62600]: DEBUG nova.compute.manager [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 665.955678] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 665.956050] env[62600]: DEBUG oslo_concurrency.lockutils [req-b0750c52-1e37-49f4-bd4c-85d1dcf910d0 req-b1011878-95c5-4fcc-929e-e12ec2ab6092 service nova] Acquired lock "refresh_cache-09ccf23b-4763-47e1-afad-59e2b06b2d9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.960303] env[62600]: DEBUG nova.network.neutron [req-b0750c52-1e37-49f4-bd4c-85d1dcf910d0 req-b1011878-95c5-4fcc-929e-e12ec2ab6092 service nova] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Refreshing network info cache for port 556c03da-1c62-48dc-a7f4-d107804f7a59 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 665.961653] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-164ee81e-4970-4313-9ca9-800b334ed622 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.976376] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7b4a2d-59b0-4d4d-9315-49c048b31c71 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.999767] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 09ccf23b-4763-47e1-afad-59e2b06b2d9e could not be found. [ 666.000066] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 666.000464] env[62600]: INFO nova.compute.manager [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 666.000529] env[62600]: DEBUG oslo.service.loopingcall [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 666.001083] env[62600]: DEBUG nova.compute.manager [-] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 666.001083] env[62600]: DEBUG nova.network.neutron [-] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 666.019946] env[62600]: DEBUG nova.network.neutron [-] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 666.156601] env[62600]: DEBUG oslo_concurrency.lockutils [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.562s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.157228] env[62600]: DEBUG nova.compute.manager [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 666.160030] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.889s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.164710] env[62600]: INFO nova.compute.claims [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 666.486878] env[62600]: DEBUG nova.network.neutron [req-b0750c52-1e37-49f4-bd4c-85d1dcf910d0 req-b1011878-95c5-4fcc-929e-e12ec2ab6092 service nova] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 666.524871] env[62600]: DEBUG nova.network.neutron [-] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.565828] env[62600]: DEBUG nova.network.neutron [req-b0750c52-1e37-49f4-bd4c-85d1dcf910d0 req-b1011878-95c5-4fcc-929e-e12ec2ab6092 service nova] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.669466] env[62600]: DEBUG nova.compute.utils [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 666.676015] env[62600]: DEBUG nova.compute.manager [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 666.677345] env[62600]: DEBUG nova.network.neutron [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 666.745267] env[62600]: DEBUG nova.policy [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c70133d8de0c4dd68bb385cdbff07a34', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '51765b8a410f4e7ea22481183447e8a9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 666.979129] env[62600]: INFO nova.scheduler.client.report [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Deleted allocations for instance 2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a [ 667.027752] env[62600]: INFO nova.compute.manager [-] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Took 1.03 seconds to deallocate network for instance. [ 667.030113] env[62600]: DEBUG nova.compute.claims [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 667.031253] env[62600]: DEBUG oslo_concurrency.lockutils [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.047485] env[62600]: DEBUG nova.network.neutron [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Successfully created port: c13d7fb7-a219-4c26-8958-fc34681c7277 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 667.072452] env[62600]: DEBUG oslo_concurrency.lockutils [req-b0750c52-1e37-49f4-bd4c-85d1dcf910d0 req-b1011878-95c5-4fcc-929e-e12ec2ab6092 service nova] Releasing lock "refresh_cache-09ccf23b-4763-47e1-afad-59e2b06b2d9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.072869] env[62600]: DEBUG nova.compute.manager [req-b0750c52-1e37-49f4-bd4c-85d1dcf910d0 req-b1011878-95c5-4fcc-929e-e12ec2ab6092 service nova] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Received event network-vif-deleted-556c03da-1c62-48dc-a7f4-d107804f7a59 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 667.094616] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Acquiring lock "dcbe8e2d-6ae9-465f-8394-3978ee61a15b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.094988] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Lock "dcbe8e2d-6ae9-465f-8394-3978ee61a15b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.176883] env[62600]: DEBUG nova.compute.manager [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 667.493221] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad10af17-8362-4fda-bc3e-f17b809c412e tempest-ServersWithSpecificFlavorTestJSON-1613978993 tempest-ServersWithSpecificFlavorTestJSON-1613978993-project-member] Lock "2e7bd3e1-6b34-4d3f-aff3-d83ffaf6317a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.038s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.695607] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b84c18-f0a2-46e7-bb4c-e2f7ba964500 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.704037] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a188565-7a61-45f6-b922-39f6dd5e2358 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.733384] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a86a8ef-7599-4f7e-9836-4c54c052edd2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.741684] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b506164f-2d77-4fff-b142-086bc784f3e2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.756848] env[62600]: DEBUG nova.compute.provider_tree [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.996962] env[62600]: DEBUG nova.compute.manager [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 668.076313] env[62600]: ERROR nova.compute.manager [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c13d7fb7-a219-4c26-8958-fc34681c7277, please check neutron logs for more information. [ 668.076313] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 668.076313] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.076313] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 668.076313] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 668.076313] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 668.076313] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 668.076313] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 668.076313] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.076313] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 668.076313] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.076313] env[62600]: ERROR nova.compute.manager raise self.value [ 668.076313] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 668.076313] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 668.076313] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.076313] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 668.077087] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.077087] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 668.077087] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c13d7fb7-a219-4c26-8958-fc34681c7277, please check neutron logs for more information. [ 668.077087] env[62600]: ERROR nova.compute.manager [ 668.077087] env[62600]: Traceback (most recent call last): [ 668.077087] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 668.077087] env[62600]: listener.cb(fileno) [ 668.077087] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.077087] env[62600]: result = function(*args, **kwargs) [ 668.077087] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 668.077087] env[62600]: return func(*args, **kwargs) [ 668.077087] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.077087] env[62600]: raise e [ 668.077087] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.077087] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 668.077087] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 668.077087] env[62600]: created_port_ids = self._update_ports_for_instance( [ 668.077087] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 668.077087] env[62600]: with excutils.save_and_reraise_exception(): [ 668.077087] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.077087] env[62600]: self.force_reraise() [ 668.077087] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.077087] env[62600]: raise self.value [ 668.077087] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 668.077087] env[62600]: updated_port = self._update_port( [ 668.077087] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.077087] env[62600]: _ensure_no_port_binding_failure(port) [ 668.077087] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.077087] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 668.077724] env[62600]: nova.exception.PortBindingFailed: Binding failed for port c13d7fb7-a219-4c26-8958-fc34681c7277, please check neutron logs for more information. [ 668.077724] env[62600]: Removing descriptor: 16 [ 668.131351] env[62600]: DEBUG nova.compute.manager [req-5230b5e8-0b5e-419e-ac72-00ab8aa7da24 req-8d8d7b00-6acc-4660-a86f-1d29b3009642 service nova] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Received event network-changed-c13d7fb7-a219-4c26-8958-fc34681c7277 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 668.131472] env[62600]: DEBUG nova.compute.manager [req-5230b5e8-0b5e-419e-ac72-00ab8aa7da24 req-8d8d7b00-6acc-4660-a86f-1d29b3009642 service nova] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Refreshing instance network info cache due to event network-changed-c13d7fb7-a219-4c26-8958-fc34681c7277. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 668.131686] env[62600]: DEBUG oslo_concurrency.lockutils [req-5230b5e8-0b5e-419e-ac72-00ab8aa7da24 req-8d8d7b00-6acc-4660-a86f-1d29b3009642 service nova] Acquiring lock "refresh_cache-60db0e43-692b-4449-9b44-badf3053429d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.131826] env[62600]: DEBUG oslo_concurrency.lockutils [req-5230b5e8-0b5e-419e-ac72-00ab8aa7da24 req-8d8d7b00-6acc-4660-a86f-1d29b3009642 service nova] Acquired lock "refresh_cache-60db0e43-692b-4449-9b44-badf3053429d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.131980] env[62600]: DEBUG nova.network.neutron [req-5230b5e8-0b5e-419e-ac72-00ab8aa7da24 req-8d8d7b00-6acc-4660-a86f-1d29b3009642 service nova] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Refreshing network info cache for port c13d7fb7-a219-4c26-8958-fc34681c7277 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 668.192538] env[62600]: DEBUG nova.compute.manager [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 668.220859] env[62600]: DEBUG nova.virt.hardware [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 668.221043] env[62600]: DEBUG nova.virt.hardware [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 668.221165] env[62600]: DEBUG nova.virt.hardware [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 668.221354] env[62600]: DEBUG nova.virt.hardware [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 668.221499] env[62600]: DEBUG nova.virt.hardware [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 668.221645] env[62600]: DEBUG nova.virt.hardware [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 668.221852] env[62600]: DEBUG nova.virt.hardware [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 668.221998] env[62600]: DEBUG nova.virt.hardware [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 668.222175] env[62600]: DEBUG nova.virt.hardware [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 668.222336] env[62600]: DEBUG nova.virt.hardware [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 668.222502] env[62600]: DEBUG nova.virt.hardware [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 668.223416] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edf61c64-5fac-48f5-a996-6bbeaf9a04b7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.231608] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5669d74-dc27-49c9-80c5-6dba8a2e3ce9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.247025] env[62600]: ERROR nova.compute.manager [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c13d7fb7-a219-4c26-8958-fc34681c7277, please check neutron logs for more information. [ 668.247025] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] Traceback (most recent call last): [ 668.247025] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 668.247025] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] yield resources [ 668.247025] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 668.247025] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] self.driver.spawn(context, instance, image_meta, [ 668.247025] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 668.247025] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 668.247025] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 668.247025] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] vm_ref = self.build_virtual_machine(instance, [ 668.247025] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 668.247420] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] vif_infos = vmwarevif.get_vif_info(self._session, [ 668.247420] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 668.247420] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] for vif in network_info: [ 668.247420] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 668.247420] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] return self._sync_wrapper(fn, *args, **kwargs) [ 668.247420] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 668.247420] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] self.wait() [ 668.247420] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 668.247420] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] self[:] = self._gt.wait() [ 668.247420] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 668.247420] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] return self._exit_event.wait() [ 668.247420] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 668.247420] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] current.throw(*self._exc) [ 668.247787] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.247787] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] result = function(*args, **kwargs) [ 668.247787] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 668.247787] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] return func(*args, **kwargs) [ 668.247787] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.247787] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] raise e [ 668.247787] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.247787] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] nwinfo = self.network_api.allocate_for_instance( [ 668.247787] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 668.247787] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] created_port_ids = self._update_ports_for_instance( [ 668.247787] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 668.247787] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] with excutils.save_and_reraise_exception(): [ 668.247787] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.248162] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] self.force_reraise() [ 668.248162] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.248162] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] raise self.value [ 668.248162] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 668.248162] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] updated_port = self._update_port( [ 668.248162] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.248162] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] _ensure_no_port_binding_failure(port) [ 668.248162] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.248162] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] raise exception.PortBindingFailed(port_id=port['id']) [ 668.248162] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] nova.exception.PortBindingFailed: Binding failed for port c13d7fb7-a219-4c26-8958-fc34681c7277, please check neutron logs for more information. [ 668.248162] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] [ 668.248162] env[62600]: INFO nova.compute.manager [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Terminating instance [ 668.249063] env[62600]: DEBUG oslo_concurrency.lockutils [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Acquiring lock "refresh_cache-60db0e43-692b-4449-9b44-badf3053429d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.259816] env[62600]: DEBUG nova.scheduler.client.report [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.520584] env[62600]: DEBUG oslo_concurrency.lockutils [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.652367] env[62600]: DEBUG nova.network.neutron [req-5230b5e8-0b5e-419e-ac72-00ab8aa7da24 req-8d8d7b00-6acc-4660-a86f-1d29b3009642 service nova] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 668.756177] env[62600]: DEBUG nova.network.neutron [req-5230b5e8-0b5e-419e-ac72-00ab8aa7da24 req-8d8d7b00-6acc-4660-a86f-1d29b3009642 service nova] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.764780] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.605s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.765489] env[62600]: DEBUG nova.compute.manager [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 668.768259] env[62600]: DEBUG oslo_concurrency.lockutils [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.663s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.261999] env[62600]: DEBUG oslo_concurrency.lockutils [req-5230b5e8-0b5e-419e-ac72-00ab8aa7da24 req-8d8d7b00-6acc-4660-a86f-1d29b3009642 service nova] Releasing lock "refresh_cache-60db0e43-692b-4449-9b44-badf3053429d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.262495] env[62600]: DEBUG oslo_concurrency.lockutils [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Acquired lock "refresh_cache-60db0e43-692b-4449-9b44-badf3053429d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.262681] env[62600]: DEBUG nova.network.neutron [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 669.272732] env[62600]: DEBUG nova.compute.utils [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 669.280211] env[62600]: DEBUG nova.compute.manager [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 669.280211] env[62600]: DEBUG nova.network.neutron [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 669.339100] env[62600]: DEBUG nova.policy [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '99e3acf8403b4db19c846f926c365d7d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3fba72f88d0c430297d71f4b1914b1be', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 669.614695] env[62600]: DEBUG nova.network.neutron [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Successfully created port: c8f65221-5a81-4065-b966-78f0d85241ce {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 669.668486] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11641c12-6c2e-4737-b75e-aaf8e9b37a3c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.676625] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e95717-54af-4e11-82a5-a842307771ce {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.715714] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc51f6ee-0086-4212-a69e-2866593e07fc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.724147] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86b61ad5-588d-431d-b785-5ac52eb3c31e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.738811] env[62600]: DEBUG nova.compute.provider_tree [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.779940] env[62600]: DEBUG nova.compute.manager [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 669.793316] env[62600]: DEBUG nova.network.neutron [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 669.916637] env[62600]: DEBUG nova.network.neutron [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.163019] env[62600]: DEBUG nova.compute.manager [req-09428d85-52e5-44c8-9dc6-e1dba87bd5f7 req-71009f45-6585-4eb4-9586-7ee417a41b03 service nova] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Received event network-vif-deleted-c13d7fb7-a219-4c26-8958-fc34681c7277 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 670.243631] env[62600]: DEBUG nova.scheduler.client.report [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 670.419371] env[62600]: DEBUG oslo_concurrency.lockutils [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Releasing lock "refresh_cache-60db0e43-692b-4449-9b44-badf3053429d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.419804] env[62600]: DEBUG nova.compute.manager [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 670.419997] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 670.420309] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c0228ef3-07b3-4050-a7cf-e4189381d6a1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.430101] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fdf4f49-398d-4158-94f7-e4a66c05cadf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.454148] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 60db0e43-692b-4449-9b44-badf3053429d could not be found. [ 670.454345] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 670.454524] env[62600]: INFO nova.compute.manager [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 670.454761] env[62600]: DEBUG oslo.service.loopingcall [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 670.454975] env[62600]: DEBUG nova.compute.manager [-] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 670.455073] env[62600]: DEBUG nova.network.neutron [-] [instance: 60db0e43-692b-4449-9b44-badf3053429d] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 670.475573] env[62600]: DEBUG nova.network.neutron [-] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.565356] env[62600]: ERROR nova.compute.manager [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c8f65221-5a81-4065-b966-78f0d85241ce, please check neutron logs for more information. [ 670.565356] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 670.565356] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.565356] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 670.565356] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 670.565356] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 670.565356] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 670.565356] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 670.565356] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.565356] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 670.565356] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.565356] env[62600]: ERROR nova.compute.manager raise self.value [ 670.565356] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 670.565356] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 670.565356] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.565356] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 670.565835] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.565835] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 670.565835] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c8f65221-5a81-4065-b966-78f0d85241ce, please check neutron logs for more information. [ 670.565835] env[62600]: ERROR nova.compute.manager [ 670.565835] env[62600]: Traceback (most recent call last): [ 670.565835] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 670.565835] env[62600]: listener.cb(fileno) [ 670.565835] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.565835] env[62600]: result = function(*args, **kwargs) [ 670.565835] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 670.565835] env[62600]: return func(*args, **kwargs) [ 670.565835] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.565835] env[62600]: raise e [ 670.565835] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.565835] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 670.565835] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 670.565835] env[62600]: created_port_ids = self._update_ports_for_instance( [ 670.565835] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 670.565835] env[62600]: with excutils.save_and_reraise_exception(): [ 670.565835] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.565835] env[62600]: self.force_reraise() [ 670.565835] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.565835] env[62600]: raise self.value [ 670.565835] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 670.565835] env[62600]: updated_port = self._update_port( [ 670.565835] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.565835] env[62600]: _ensure_no_port_binding_failure(port) [ 670.565835] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.565835] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 670.566687] env[62600]: nova.exception.PortBindingFailed: Binding failed for port c8f65221-5a81-4065-b966-78f0d85241ce, please check neutron logs for more information. [ 670.566687] env[62600]: Removing descriptor: 16 [ 670.749454] env[62600]: DEBUG oslo_concurrency.lockutils [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.981s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.750141] env[62600]: ERROR nova.compute.manager [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9cf2f0af-7904-459f-9a8e-a9244e7a5fc7, please check neutron logs for more information. [ 670.750141] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Traceback (most recent call last): [ 670.750141] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 670.750141] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] self.driver.spawn(context, instance, image_meta, [ 670.750141] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 670.750141] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.750141] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.750141] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] vm_ref = self.build_virtual_machine(instance, [ 670.750141] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.750141] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.750141] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.750623] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] for vif in network_info: [ 670.750623] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.750623] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] return self._sync_wrapper(fn, *args, **kwargs) [ 670.750623] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.750623] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] self.wait() [ 670.750623] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.750623] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] self[:] = self._gt.wait() [ 670.750623] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.750623] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] return self._exit_event.wait() [ 670.750623] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 670.750623] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] result = hub.switch() [ 670.750623] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 670.750623] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] return self.greenlet.switch() [ 670.751213] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.751213] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] result = function(*args, **kwargs) [ 670.751213] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 670.751213] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] return func(*args, **kwargs) [ 670.751213] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.751213] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] raise e [ 670.751213] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.751213] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] nwinfo = self.network_api.allocate_for_instance( [ 670.751213] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 670.751213] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] created_port_ids = self._update_ports_for_instance( [ 670.751213] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 670.751213] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] with excutils.save_and_reraise_exception(): [ 670.751213] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.751565] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] self.force_reraise() [ 670.751565] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.751565] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] raise self.value [ 670.751565] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 670.751565] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] updated_port = self._update_port( [ 670.751565] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.751565] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] _ensure_no_port_binding_failure(port) [ 670.751565] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.751565] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] raise exception.PortBindingFailed(port_id=port['id']) [ 670.751565] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] nova.exception.PortBindingFailed: Binding failed for port 9cf2f0af-7904-459f-9a8e-a9244e7a5fc7, please check neutron logs for more information. [ 670.751565] env[62600]: ERROR nova.compute.manager [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] [ 670.751850] env[62600]: DEBUG nova.compute.utils [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Binding failed for port 9cf2f0af-7904-459f-9a8e-a9244e7a5fc7, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 670.752364] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.261s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.754046] env[62600]: INFO nova.compute.claims [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 670.757957] env[62600]: DEBUG nova.compute.manager [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Build of instance 222a94b5-7841-490e-9b27-d3f703e6794f was re-scheduled: Binding failed for port 9cf2f0af-7904-459f-9a8e-a9244e7a5fc7, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 670.757957] env[62600]: DEBUG nova.compute.manager [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 670.757957] env[62600]: DEBUG oslo_concurrency.lockutils [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquiring lock "refresh_cache-222a94b5-7841-490e-9b27-d3f703e6794f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.757957] env[62600]: DEBUG oslo_concurrency.lockutils [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquired lock "refresh_cache-222a94b5-7841-490e-9b27-d3f703e6794f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.758257] env[62600]: DEBUG nova.network.neutron [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 670.789048] env[62600]: DEBUG nova.compute.manager [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 670.826031] env[62600]: DEBUG nova.virt.hardware [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 670.826806] env[62600]: DEBUG nova.virt.hardware [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 670.826806] env[62600]: DEBUG nova.virt.hardware [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 670.826806] env[62600]: DEBUG nova.virt.hardware [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 670.826944] env[62600]: DEBUG nova.virt.hardware [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 670.826975] env[62600]: DEBUG nova.virt.hardware [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 670.827258] env[62600]: DEBUG nova.virt.hardware [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 670.827427] env[62600]: DEBUG nova.virt.hardware [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 670.827595] env[62600]: DEBUG nova.virt.hardware [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 670.827756] env[62600]: DEBUG nova.virt.hardware [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 670.828057] env[62600]: DEBUG nova.virt.hardware [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 670.829517] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61abad1-0212-4cf6-abf0-6df4d78a139d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.837632] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c56751c5-9cc5-4132-a380-d5fbabb58ef9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.852463] env[62600]: ERROR nova.compute.manager [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c8f65221-5a81-4065-b966-78f0d85241ce, please check neutron logs for more information. [ 670.852463] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Traceback (most recent call last): [ 670.852463] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 670.852463] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] yield resources [ 670.852463] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 670.852463] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] self.driver.spawn(context, instance, image_meta, [ 670.852463] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 670.852463] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.852463] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.852463] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] vm_ref = self.build_virtual_machine(instance, [ 670.852463] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.852789] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.852789] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.852789] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] for vif in network_info: [ 670.852789] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.852789] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] return self._sync_wrapper(fn, *args, **kwargs) [ 670.852789] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.852789] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] self.wait() [ 670.852789] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.852789] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] self[:] = self._gt.wait() [ 670.852789] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.852789] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] return self._exit_event.wait() [ 670.852789] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 670.852789] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] current.throw(*self._exc) [ 670.853101] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.853101] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] result = function(*args, **kwargs) [ 670.853101] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 670.853101] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] return func(*args, **kwargs) [ 670.853101] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.853101] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] raise e [ 670.853101] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.853101] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] nwinfo = self.network_api.allocate_for_instance( [ 670.853101] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 670.853101] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] created_port_ids = self._update_ports_for_instance( [ 670.853101] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 670.853101] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] with excutils.save_and_reraise_exception(): [ 670.853101] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.853484] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] self.force_reraise() [ 670.853484] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.853484] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] raise self.value [ 670.853484] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 670.853484] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] updated_port = self._update_port( [ 670.853484] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.853484] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] _ensure_no_port_binding_failure(port) [ 670.853484] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.853484] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] raise exception.PortBindingFailed(port_id=port['id']) [ 670.853484] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] nova.exception.PortBindingFailed: Binding failed for port c8f65221-5a81-4065-b966-78f0d85241ce, please check neutron logs for more information. [ 670.853484] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] [ 670.853484] env[62600]: INFO nova.compute.manager [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Terminating instance [ 670.855180] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Acquiring lock "refresh_cache-9fd421c6-d068-4a81-a042-2d1f23fb9e1a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.855180] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Acquired lock "refresh_cache-9fd421c6-d068-4a81-a042-2d1f23fb9e1a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.855180] env[62600]: DEBUG nova.network.neutron [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 670.977996] env[62600]: DEBUG nova.network.neutron [-] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.284332] env[62600]: DEBUG nova.network.neutron [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.474187] env[62600]: DEBUG nova.network.neutron [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.483445] env[62600]: INFO nova.compute.manager [-] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Took 1.03 seconds to deallocate network for instance. [ 671.487032] env[62600]: DEBUG nova.compute.claims [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 671.487219] env[62600]: DEBUG oslo_concurrency.lockutils [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.610091] env[62600]: DEBUG nova.network.neutron [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.617247] env[62600]: DEBUG nova.network.neutron [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.112555] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Releasing lock "refresh_cache-9fd421c6-d068-4a81-a042-2d1f23fb9e1a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.112993] env[62600]: DEBUG nova.compute.manager [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 672.113211] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 672.113508] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e14bac45-aa4b-458f-aa22-64ca4a2b1d01 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.124021] env[62600]: DEBUG oslo_concurrency.lockutils [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Releasing lock "refresh_cache-222a94b5-7841-490e-9b27-d3f703e6794f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.124021] env[62600]: DEBUG nova.compute.manager [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 672.124021] env[62600]: DEBUG nova.compute.manager [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.124021] env[62600]: DEBUG nova.network.neutron [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 672.131416] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d0b031-51d5-40b4-882c-47773c3210bc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.145556] env[62600]: DEBUG nova.network.neutron [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.163359] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9fd421c6-d068-4a81-a042-2d1f23fb9e1a could not be found. [ 672.163584] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 672.163760] env[62600]: INFO nova.compute.manager [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Took 0.05 seconds to destroy the instance on the hypervisor. [ 672.164031] env[62600]: DEBUG oslo.service.loopingcall [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 672.164236] env[62600]: DEBUG nova.compute.manager [-] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.164342] env[62600]: DEBUG nova.network.neutron [-] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 672.168283] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b3d57c5-97e8-4715-a752-21f7d5e6f467 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.176541] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a22fef-8a95-4125-83ea-3438308a51ff {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.209368] env[62600]: DEBUG nova.network.neutron [-] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.211995] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-516bf1d5-5e1a-4228-bd4d-42c1749bcc3d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.215445] env[62600]: DEBUG nova.compute.manager [req-74d90b96-c042-40bf-8028-0cf5a631dc2f req-9860d4d2-f90a-4104-8a8d-1d348c29213c service nova] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Received event network-changed-c8f65221-5a81-4065-b966-78f0d85241ce {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 672.215762] env[62600]: DEBUG nova.compute.manager [req-74d90b96-c042-40bf-8028-0cf5a631dc2f req-9860d4d2-f90a-4104-8a8d-1d348c29213c service nova] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Refreshing instance network info cache due to event network-changed-c8f65221-5a81-4065-b966-78f0d85241ce. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 672.216076] env[62600]: DEBUG oslo_concurrency.lockutils [req-74d90b96-c042-40bf-8028-0cf5a631dc2f req-9860d4d2-f90a-4104-8a8d-1d348c29213c service nova] Acquiring lock "refresh_cache-9fd421c6-d068-4a81-a042-2d1f23fb9e1a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.216268] env[62600]: DEBUG oslo_concurrency.lockutils [req-74d90b96-c042-40bf-8028-0cf5a631dc2f req-9860d4d2-f90a-4104-8a8d-1d348c29213c service nova] Acquired lock "refresh_cache-9fd421c6-d068-4a81-a042-2d1f23fb9e1a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.216463] env[62600]: DEBUG nova.network.neutron [req-74d90b96-c042-40bf-8028-0cf5a631dc2f req-9860d4d2-f90a-4104-8a8d-1d348c29213c service nova] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Refreshing network info cache for port c8f65221-5a81-4065-b966-78f0d85241ce {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 672.223462] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e37a73-d41b-404d-b185-00589aa8d815 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.238992] env[62600]: DEBUG nova.compute.provider_tree [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.653095] env[62600]: DEBUG nova.network.neutron [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.720320] env[62600]: DEBUG nova.network.neutron [-] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.742188] env[62600]: DEBUG nova.scheduler.client.report [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 672.750424] env[62600]: DEBUG nova.network.neutron [req-74d90b96-c042-40bf-8028-0cf5a631dc2f req-9860d4d2-f90a-4104-8a8d-1d348c29213c service nova] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.833178] env[62600]: DEBUG nova.network.neutron [req-74d90b96-c042-40bf-8028-0cf5a631dc2f req-9860d4d2-f90a-4104-8a8d-1d348c29213c service nova] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.158471] env[62600]: INFO nova.compute.manager [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 222a94b5-7841-490e-9b27-d3f703e6794f] Took 1.04 seconds to deallocate network for instance. [ 673.223990] env[62600]: INFO nova.compute.manager [-] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Took 1.06 seconds to deallocate network for instance. [ 673.229189] env[62600]: DEBUG nova.compute.claims [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 673.229189] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.252768] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquiring lock "b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.252912] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.254816] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.502s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.255325] env[62600]: DEBUG nova.compute.manager [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 673.258410] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.970s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.259722] env[62600]: INFO nova.compute.claims [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 673.336085] env[62600]: DEBUG oslo_concurrency.lockutils [req-74d90b96-c042-40bf-8028-0cf5a631dc2f req-9860d4d2-f90a-4104-8a8d-1d348c29213c service nova] Releasing lock "refresh_cache-9fd421c6-d068-4a81-a042-2d1f23fb9e1a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.336355] env[62600]: DEBUG nova.compute.manager [req-74d90b96-c042-40bf-8028-0cf5a631dc2f req-9860d4d2-f90a-4104-8a8d-1d348c29213c service nova] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Received event network-vif-deleted-c8f65221-5a81-4065-b966-78f0d85241ce {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 673.760131] env[62600]: DEBUG nova.compute.utils [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 673.761617] env[62600]: DEBUG nova.compute.manager [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 673.761617] env[62600]: DEBUG nova.network.neutron [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 673.782975] env[62600]: DEBUG oslo_concurrency.lockutils [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquiring lock "d34c098d-a827-4dbd-96e4-e27d9d56b847" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.782975] env[62600]: DEBUG oslo_concurrency.lockutils [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "d34c098d-a827-4dbd-96e4-e27d9d56b847" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.818021] env[62600]: DEBUG nova.policy [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3c4a5f4b15bd4c689e7f6cd868787c63', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ce1553057d14d86b5344158a6148dcd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 674.129871] env[62600]: DEBUG nova.network.neutron [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Successfully created port: 8f0a9399-ca7f-4eb3-b9f3-c5a1b6a20b46 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 674.191649] env[62600]: INFO nova.scheduler.client.report [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Deleted allocations for instance 222a94b5-7841-490e-9b27-d3f703e6794f [ 674.265266] env[62600]: DEBUG nova.compute.manager [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 674.697095] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e31e2c-0d9b-4832-95ab-c38e71fb5773 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.704018] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72c93543-28fc-456e-8bb8-49f73def8c87 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.706773] env[62600]: DEBUG oslo_concurrency.lockutils [None req-005819b1-8395-466a-a709-83ebebe8c7a0 tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Lock "222a94b5-7841-490e-9b27-d3f703e6794f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.785s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.738408] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78de48e-85c6-4a79-b81e-9adacbfd15e4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.746774] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c55b6f98-8a8c-4a31-9293-8aad12f55901 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.763549] env[62600]: DEBUG nova.compute.provider_tree [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.842357] env[62600]: DEBUG nova.compute.manager [req-6800b6ff-6e46-4e87-8392-cad4242aa0f1 req-47dc3fe7-aa4a-4411-908c-5b1ee700f274 service nova] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Received event network-changed-8f0a9399-ca7f-4eb3-b9f3-c5a1b6a20b46 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 674.842555] env[62600]: DEBUG nova.compute.manager [req-6800b6ff-6e46-4e87-8392-cad4242aa0f1 req-47dc3fe7-aa4a-4411-908c-5b1ee700f274 service nova] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Refreshing instance network info cache due to event network-changed-8f0a9399-ca7f-4eb3-b9f3-c5a1b6a20b46. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 674.842770] env[62600]: DEBUG oslo_concurrency.lockutils [req-6800b6ff-6e46-4e87-8392-cad4242aa0f1 req-47dc3fe7-aa4a-4411-908c-5b1ee700f274 service nova] Acquiring lock "refresh_cache-e8cbb437-0ace-4af3-a1d9-f96eeeea8504" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.842914] env[62600]: DEBUG oslo_concurrency.lockutils [req-6800b6ff-6e46-4e87-8392-cad4242aa0f1 req-47dc3fe7-aa4a-4411-908c-5b1ee700f274 service nova] Acquired lock "refresh_cache-e8cbb437-0ace-4af3-a1d9-f96eeeea8504" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.843089] env[62600]: DEBUG nova.network.neutron [req-6800b6ff-6e46-4e87-8392-cad4242aa0f1 req-47dc3fe7-aa4a-4411-908c-5b1ee700f274 service nova] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Refreshing network info cache for port 8f0a9399-ca7f-4eb3-b9f3-c5a1b6a20b46 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 675.024327] env[62600]: ERROR nova.compute.manager [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8f0a9399-ca7f-4eb3-b9f3-c5a1b6a20b46, please check neutron logs for more information. [ 675.024327] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 675.024327] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.024327] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 675.024327] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.024327] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 675.024327] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.024327] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 675.024327] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.024327] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 675.024327] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.024327] env[62600]: ERROR nova.compute.manager raise self.value [ 675.024327] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.024327] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 675.024327] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.024327] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 675.024790] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.024790] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 675.024790] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8f0a9399-ca7f-4eb3-b9f3-c5a1b6a20b46, please check neutron logs for more information. [ 675.024790] env[62600]: ERROR nova.compute.manager [ 675.024790] env[62600]: Traceback (most recent call last): [ 675.024790] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 675.024790] env[62600]: listener.cb(fileno) [ 675.024790] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.024790] env[62600]: result = function(*args, **kwargs) [ 675.024790] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.024790] env[62600]: return func(*args, **kwargs) [ 675.024790] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.024790] env[62600]: raise e [ 675.024790] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.024790] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 675.024790] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.024790] env[62600]: created_port_ids = self._update_ports_for_instance( [ 675.024790] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.024790] env[62600]: with excutils.save_and_reraise_exception(): [ 675.024790] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.024790] env[62600]: self.force_reraise() [ 675.024790] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.024790] env[62600]: raise self.value [ 675.024790] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.024790] env[62600]: updated_port = self._update_port( [ 675.024790] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.024790] env[62600]: _ensure_no_port_binding_failure(port) [ 675.024790] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.024790] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 675.025569] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 8f0a9399-ca7f-4eb3-b9f3-c5a1b6a20b46, please check neutron logs for more information. [ 675.025569] env[62600]: Removing descriptor: 16 [ 675.209514] env[62600]: DEBUG nova.compute.manager [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 675.268638] env[62600]: DEBUG nova.scheduler.client.report [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 675.277176] env[62600]: DEBUG nova.compute.manager [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 675.305275] env[62600]: DEBUG nova.virt.hardware [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 675.305396] env[62600]: DEBUG nova.virt.hardware [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 675.305512] env[62600]: DEBUG nova.virt.hardware [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 675.305693] env[62600]: DEBUG nova.virt.hardware [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 675.305840] env[62600]: DEBUG nova.virt.hardware [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 675.306012] env[62600]: DEBUG nova.virt.hardware [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 675.310381] env[62600]: DEBUG nova.virt.hardware [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 675.310564] env[62600]: DEBUG nova.virt.hardware [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 675.310741] env[62600]: DEBUG nova.virt.hardware [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 675.310923] env[62600]: DEBUG nova.virt.hardware [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 675.311099] env[62600]: DEBUG nova.virt.hardware [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 675.312397] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ff5fc1-98f4-451b-bf50-69427e42774a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.323649] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30f2124-28c8-4f68-b236-c9c45b3df988 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.338766] env[62600]: ERROR nova.compute.manager [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8f0a9399-ca7f-4eb3-b9f3-c5a1b6a20b46, please check neutron logs for more information. [ 675.338766] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Traceback (most recent call last): [ 675.338766] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 675.338766] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] yield resources [ 675.338766] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 675.338766] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] self.driver.spawn(context, instance, image_meta, [ 675.338766] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 675.338766] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] self._vmops.spawn(context, instance, image_meta, injected_files, [ 675.338766] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 675.338766] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] vm_ref = self.build_virtual_machine(instance, [ 675.338766] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 675.339062] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] vif_infos = vmwarevif.get_vif_info(self._session, [ 675.339062] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 675.339062] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] for vif in network_info: [ 675.339062] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 675.339062] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] return self._sync_wrapper(fn, *args, **kwargs) [ 675.339062] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 675.339062] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] self.wait() [ 675.339062] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 675.339062] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] self[:] = self._gt.wait() [ 675.339062] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 675.339062] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] return self._exit_event.wait() [ 675.339062] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 675.339062] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] current.throw(*self._exc) [ 675.339368] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.339368] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] result = function(*args, **kwargs) [ 675.339368] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 675.339368] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] return func(*args, **kwargs) [ 675.339368] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.339368] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] raise e [ 675.339368] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.339368] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] nwinfo = self.network_api.allocate_for_instance( [ 675.339368] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 675.339368] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] created_port_ids = self._update_ports_for_instance( [ 675.339368] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 675.339368] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] with excutils.save_and_reraise_exception(): [ 675.339368] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.339666] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] self.force_reraise() [ 675.339666] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.339666] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] raise self.value [ 675.339666] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 675.339666] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] updated_port = self._update_port( [ 675.339666] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.339666] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] _ensure_no_port_binding_failure(port) [ 675.339666] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.339666] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] raise exception.PortBindingFailed(port_id=port['id']) [ 675.339666] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] nova.exception.PortBindingFailed: Binding failed for port 8f0a9399-ca7f-4eb3-b9f3-c5a1b6a20b46, please check neutron logs for more information. [ 675.339666] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] [ 675.339666] env[62600]: INFO nova.compute.manager [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Terminating instance [ 675.341116] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Acquiring lock "refresh_cache-e8cbb437-0ace-4af3-a1d9-f96eeeea8504" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.362661] env[62600]: DEBUG nova.network.neutron [req-6800b6ff-6e46-4e87-8392-cad4242aa0f1 req-47dc3fe7-aa4a-4411-908c-5b1ee700f274 service nova] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 675.454186] env[62600]: DEBUG nova.network.neutron [req-6800b6ff-6e46-4e87-8392-cad4242aa0f1 req-47dc3fe7-aa4a-4411-908c-5b1ee700f274 service nova] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.737795] env[62600]: DEBUG oslo_concurrency.lockutils [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.771148] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.513s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.771686] env[62600]: DEBUG nova.compute.manager [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 675.774698] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.613s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.774698] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.775178] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62600) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 675.775178] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.417s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.776632] env[62600]: INFO nova.compute.claims [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 675.779874] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99cf94d2-9528-4a27-8adc-98905df4b85f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.789880] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d9c7cca-c877-43c3-98eb-db9a9cb4f9cc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.805791] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7247cc05-532a-4dfd-9259-3a9d57dc22df {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.812291] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c264cc53-8b8c-492b-8b83-42854a67ff1b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.844022] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181466MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=62600) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 675.844022] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.957526] env[62600]: DEBUG oslo_concurrency.lockutils [req-6800b6ff-6e46-4e87-8392-cad4242aa0f1 req-47dc3fe7-aa4a-4411-908c-5b1ee700f274 service nova] Releasing lock "refresh_cache-e8cbb437-0ace-4af3-a1d9-f96eeeea8504" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.958138] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Acquired lock "refresh_cache-e8cbb437-0ace-4af3-a1d9-f96eeeea8504" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.958343] env[62600]: DEBUG nova.network.neutron [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 676.280903] env[62600]: DEBUG nova.compute.utils [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 676.282342] env[62600]: DEBUG nova.compute.manager [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 676.282514] env[62600]: DEBUG nova.network.neutron [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 676.368052] env[62600]: DEBUG nova.policy [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '94d9eac6ec93491db7f71ec82c30e074', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab845bd31c174ce0aad988dc61e5368e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 676.479601] env[62600]: DEBUG nova.network.neutron [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.604887] env[62600]: DEBUG nova.network.neutron [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.676084] env[62600]: DEBUG nova.network.neutron [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Successfully created port: 15dceed5-8643-4f6b-ba3c-0ec3b34f0684 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 676.789822] env[62600]: DEBUG nova.compute.manager [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 677.041286] env[62600]: DEBUG nova.compute.manager [req-fa14a1da-6f6d-44f9-ac70-c175272a60f1 req-1b343ffb-113c-4792-9647-b5fafee16ea1 service nova] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Received event network-vif-deleted-8f0a9399-ca7f-4eb3-b9f3-c5a1b6a20b46 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 677.109425] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Releasing lock "refresh_cache-e8cbb437-0ace-4af3-a1d9-f96eeeea8504" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.110484] env[62600]: DEBUG nova.compute.manager [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 677.111592] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 677.111631] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b8e4fb05-0cee-4860-a545-fc65fdcd46b9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.122650] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349a2d41-e44d-401b-961c-72a5d58db1eb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.155505] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e8cbb437-0ace-4af3-a1d9-f96eeeea8504 could not be found. [ 677.155739] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 677.155921] env[62600]: INFO nova.compute.manager [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Took 0.05 seconds to destroy the instance on the hypervisor. [ 677.156352] env[62600]: DEBUG oslo.service.loopingcall [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 677.158898] env[62600]: DEBUG nova.compute.manager [-] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 677.159009] env[62600]: DEBUG nova.network.neutron [-] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 677.177074] env[62600]: DEBUG nova.network.neutron [-] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.231097] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-264adebb-7cff-4d55-93f7-2dfb2f557ad3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.238920] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d962166-b75d-4827-bba6-cc300574c571 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.279175] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fdd63de-2f2c-483f-84ff-a3116632c35b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.287398] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd6ffe3a-16b6-47f8-a0b5-84a436dff485 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.303880] env[62600]: DEBUG nova.compute.provider_tree [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 677.560579] env[62600]: ERROR nova.compute.manager [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 15dceed5-8643-4f6b-ba3c-0ec3b34f0684, please check neutron logs for more information. [ 677.560579] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 677.560579] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.560579] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 677.560579] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.560579] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 677.560579] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.560579] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 677.560579] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.560579] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 677.560579] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.560579] env[62600]: ERROR nova.compute.manager raise self.value [ 677.560579] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.560579] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 677.560579] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.560579] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 677.561220] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.561220] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 677.561220] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 15dceed5-8643-4f6b-ba3c-0ec3b34f0684, please check neutron logs for more information. [ 677.561220] env[62600]: ERROR nova.compute.manager [ 677.561220] env[62600]: Traceback (most recent call last): [ 677.561220] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 677.561220] env[62600]: listener.cb(fileno) [ 677.561220] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.561220] env[62600]: result = function(*args, **kwargs) [ 677.561220] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.561220] env[62600]: return func(*args, **kwargs) [ 677.561220] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.561220] env[62600]: raise e [ 677.561220] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.561220] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 677.561220] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.561220] env[62600]: created_port_ids = self._update_ports_for_instance( [ 677.561220] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.561220] env[62600]: with excutils.save_and_reraise_exception(): [ 677.561220] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.561220] env[62600]: self.force_reraise() [ 677.561220] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.561220] env[62600]: raise self.value [ 677.561220] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.561220] env[62600]: updated_port = self._update_port( [ 677.561220] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.561220] env[62600]: _ensure_no_port_binding_failure(port) [ 677.561220] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.561220] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 677.561993] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 15dceed5-8643-4f6b-ba3c-0ec3b34f0684, please check neutron logs for more information. [ 677.561993] env[62600]: Removing descriptor: 16 [ 677.679389] env[62600]: DEBUG nova.network.neutron [-] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.806435] env[62600]: DEBUG nova.compute.manager [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 677.810576] env[62600]: DEBUG nova.scheduler.client.report [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 677.843575] env[62600]: DEBUG nova.virt.hardware [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 677.843810] env[62600]: DEBUG nova.virt.hardware [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 677.843963] env[62600]: DEBUG nova.virt.hardware [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 677.844162] env[62600]: DEBUG nova.virt.hardware [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 677.844305] env[62600]: DEBUG nova.virt.hardware [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 677.844446] env[62600]: DEBUG nova.virt.hardware [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 677.844642] env[62600]: DEBUG nova.virt.hardware [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 677.844791] env[62600]: DEBUG nova.virt.hardware [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 677.844948] env[62600]: DEBUG nova.virt.hardware [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 677.846235] env[62600]: DEBUG nova.virt.hardware [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 677.846235] env[62600]: DEBUG nova.virt.hardware [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 677.846559] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c11c2156-36b0-468e-a3f6-7bbc637800f4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.854739] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3060910-7658-40fa-9a55-532835a090d2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.870610] env[62600]: ERROR nova.compute.manager [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 15dceed5-8643-4f6b-ba3c-0ec3b34f0684, please check neutron logs for more information. [ 677.870610] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] Traceback (most recent call last): [ 677.870610] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 677.870610] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] yield resources [ 677.870610] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.870610] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] self.driver.spawn(context, instance, image_meta, [ 677.870610] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 677.870610] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.870610] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.870610] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] vm_ref = self.build_virtual_machine(instance, [ 677.870610] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.871107] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.871107] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.871107] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] for vif in network_info: [ 677.871107] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.871107] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] return self._sync_wrapper(fn, *args, **kwargs) [ 677.871107] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.871107] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] self.wait() [ 677.871107] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.871107] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] self[:] = self._gt.wait() [ 677.871107] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.871107] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] return self._exit_event.wait() [ 677.871107] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 677.871107] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] current.throw(*self._exc) [ 677.871424] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.871424] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] result = function(*args, **kwargs) [ 677.871424] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.871424] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] return func(*args, **kwargs) [ 677.871424] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.871424] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] raise e [ 677.871424] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.871424] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] nwinfo = self.network_api.allocate_for_instance( [ 677.871424] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.871424] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] created_port_ids = self._update_ports_for_instance( [ 677.871424] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.871424] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] with excutils.save_and_reraise_exception(): [ 677.871424] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.871809] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] self.force_reraise() [ 677.871809] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.871809] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] raise self.value [ 677.871809] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.871809] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] updated_port = self._update_port( [ 677.871809] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.871809] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] _ensure_no_port_binding_failure(port) [ 677.871809] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.871809] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] raise exception.PortBindingFailed(port_id=port['id']) [ 677.871809] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] nova.exception.PortBindingFailed: Binding failed for port 15dceed5-8643-4f6b-ba3c-0ec3b34f0684, please check neutron logs for more information. [ 677.871809] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] [ 677.871809] env[62600]: INFO nova.compute.manager [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Terminating instance [ 677.873838] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Acquiring lock "refresh_cache-3d19d349-e450-4307-a763-6269dcb06544" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.873838] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Acquired lock "refresh_cache-3d19d349-e450-4307-a763-6269dcb06544" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.874017] env[62600]: DEBUG nova.network.neutron [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 678.181810] env[62600]: INFO nova.compute.manager [-] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Took 1.02 seconds to deallocate network for instance. [ 678.184872] env[62600]: DEBUG nova.compute.claims [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 678.185308] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.315936] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.541s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.321160] env[62600]: DEBUG nova.compute.manager [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 678.321160] env[62600]: DEBUG oslo_concurrency.lockutils [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.011s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.322584] env[62600]: INFO nova.compute.claims [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 678.397089] env[62600]: DEBUG nova.network.neutron [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.529754] env[62600]: DEBUG nova.network.neutron [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.828515] env[62600]: DEBUG nova.compute.utils [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 678.835513] env[62600]: DEBUG nova.compute.manager [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 678.835513] env[62600]: DEBUG nova.network.neutron [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 678.884450] env[62600]: DEBUG nova.policy [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9d52716f2115422a97936f53e876dbb9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1c0c8f546674eee8993ba2adae79407', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 679.035446] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Releasing lock "refresh_cache-3d19d349-e450-4307-a763-6269dcb06544" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.035446] env[62600]: DEBUG nova.compute.manager [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 679.035446] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 679.035446] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fcd80ef0-232c-4454-a8c9-532cc531f490 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.047847] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aba196c-9890-4460-972f-0f669ff4c62b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.073242] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3d19d349-e450-4307-a763-6269dcb06544 could not be found. [ 679.073242] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 679.073242] env[62600]: INFO nova.compute.manager [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Took 0.04 seconds to destroy the instance on the hypervisor. [ 679.073242] env[62600]: DEBUG oslo.service.loopingcall [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 679.073242] env[62600]: DEBUG nova.compute.manager [-] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 679.073242] env[62600]: DEBUG nova.network.neutron [-] [instance: 3d19d349-e450-4307-a763-6269dcb06544] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 679.095498] env[62600]: DEBUG nova.network.neutron [-] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.188210] env[62600]: DEBUG nova.network.neutron [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Successfully created port: cfd31541-18d4-4711-80f4-ee76d04de9e9 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 679.334220] env[62600]: DEBUG nova.compute.manager [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 679.368950] env[62600]: DEBUG nova.compute.manager [req-2de8d5c5-ebed-4820-902e-5ff24de6b27d req-3d595eb9-7202-4275-816e-2dfa3f45e41e service nova] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Received event network-changed-15dceed5-8643-4f6b-ba3c-0ec3b34f0684 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 679.368950] env[62600]: DEBUG nova.compute.manager [req-2de8d5c5-ebed-4820-902e-5ff24de6b27d req-3d595eb9-7202-4275-816e-2dfa3f45e41e service nova] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Refreshing instance network info cache due to event network-changed-15dceed5-8643-4f6b-ba3c-0ec3b34f0684. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 679.369169] env[62600]: DEBUG oslo_concurrency.lockutils [req-2de8d5c5-ebed-4820-902e-5ff24de6b27d req-3d595eb9-7202-4275-816e-2dfa3f45e41e service nova] Acquiring lock "refresh_cache-3d19d349-e450-4307-a763-6269dcb06544" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.369316] env[62600]: DEBUG oslo_concurrency.lockutils [req-2de8d5c5-ebed-4820-902e-5ff24de6b27d req-3d595eb9-7202-4275-816e-2dfa3f45e41e service nova] Acquired lock "refresh_cache-3d19d349-e450-4307-a763-6269dcb06544" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.369528] env[62600]: DEBUG nova.network.neutron [req-2de8d5c5-ebed-4820-902e-5ff24de6b27d req-3d595eb9-7202-4275-816e-2dfa3f45e41e service nova] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Refreshing network info cache for port 15dceed5-8643-4f6b-ba3c-0ec3b34f0684 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 679.598356] env[62600]: DEBUG nova.network.neutron [-] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.682082] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquiring lock "92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.682396] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Lock "92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.004s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.790513] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-221e3336-596c-42a3-beb3-2be49291b734 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.798573] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fde9dac-7ee3-4fd7-a23e-414801aac05f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.828299] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e57736-5a22-4c97-a25d-2e9a85b3004c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.836198] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a89d5b1a-3836-4399-9614-4b4376b2a0d2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.854472] env[62600]: DEBUG nova.compute.provider_tree [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.894177] env[62600]: DEBUG nova.network.neutron [req-2de8d5c5-ebed-4820-902e-5ff24de6b27d req-3d595eb9-7202-4275-816e-2dfa3f45e41e service nova] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.035256] env[62600]: DEBUG nova.network.neutron [req-2de8d5c5-ebed-4820-902e-5ff24de6b27d req-3d595eb9-7202-4275-816e-2dfa3f45e41e service nova] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.105284] env[62600]: INFO nova.compute.manager [-] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Took 1.03 seconds to deallocate network for instance. [ 680.107185] env[62600]: DEBUG nova.compute.claims [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 680.107495] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.169574] env[62600]: ERROR nova.compute.manager [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cfd31541-18d4-4711-80f4-ee76d04de9e9, please check neutron logs for more information. [ 680.169574] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 680.169574] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.169574] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 680.169574] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.169574] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 680.169574] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.169574] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 680.169574] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.169574] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 680.169574] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.169574] env[62600]: ERROR nova.compute.manager raise self.value [ 680.169574] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.169574] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 680.169574] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.169574] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 680.169959] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.169959] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 680.169959] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cfd31541-18d4-4711-80f4-ee76d04de9e9, please check neutron logs for more information. [ 680.169959] env[62600]: ERROR nova.compute.manager [ 680.169959] env[62600]: Traceback (most recent call last): [ 680.169959] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 680.169959] env[62600]: listener.cb(fileno) [ 680.169959] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.169959] env[62600]: result = function(*args, **kwargs) [ 680.169959] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.169959] env[62600]: return func(*args, **kwargs) [ 680.169959] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.169959] env[62600]: raise e [ 680.169959] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.169959] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 680.169959] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.169959] env[62600]: created_port_ids = self._update_ports_for_instance( [ 680.169959] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.169959] env[62600]: with excutils.save_and_reraise_exception(): [ 680.169959] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.169959] env[62600]: self.force_reraise() [ 680.169959] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.169959] env[62600]: raise self.value [ 680.169959] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.169959] env[62600]: updated_port = self._update_port( [ 680.169959] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.169959] env[62600]: _ensure_no_port_binding_failure(port) [ 680.169959] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.169959] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 680.170629] env[62600]: nova.exception.PortBindingFailed: Binding failed for port cfd31541-18d4-4711-80f4-ee76d04de9e9, please check neutron logs for more information. [ 680.170629] env[62600]: Removing descriptor: 16 [ 680.357332] env[62600]: DEBUG nova.compute.manager [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 680.360399] env[62600]: DEBUG nova.scheduler.client.report [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 680.383010] env[62600]: DEBUG nova.virt.hardware [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 680.383319] env[62600]: DEBUG nova.virt.hardware [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 680.383489] env[62600]: DEBUG nova.virt.hardware [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 680.383672] env[62600]: DEBUG nova.virt.hardware [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 680.383822] env[62600]: DEBUG nova.virt.hardware [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 680.383974] env[62600]: DEBUG nova.virt.hardware [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 680.384198] env[62600]: DEBUG nova.virt.hardware [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 680.384359] env[62600]: DEBUG nova.virt.hardware [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 680.384524] env[62600]: DEBUG nova.virt.hardware [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 680.384687] env[62600]: DEBUG nova.virt.hardware [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 680.384972] env[62600]: DEBUG nova.virt.hardware [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 680.385732] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2085f093-54f0-438d-a9df-e17d4f43eb37 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.394419] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109b6a2e-63ee-46e3-875a-ae9ee43aa796 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.408651] env[62600]: ERROR nova.compute.manager [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cfd31541-18d4-4711-80f4-ee76d04de9e9, please check neutron logs for more information. [ 680.408651] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Traceback (most recent call last): [ 680.408651] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 680.408651] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] yield resources [ 680.408651] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 680.408651] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] self.driver.spawn(context, instance, image_meta, [ 680.408651] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 680.408651] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.408651] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.408651] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] vm_ref = self.build_virtual_machine(instance, [ 680.408651] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.408993] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.408993] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.408993] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] for vif in network_info: [ 680.408993] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.408993] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] return self._sync_wrapper(fn, *args, **kwargs) [ 680.408993] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.408993] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] self.wait() [ 680.408993] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.408993] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] self[:] = self._gt.wait() [ 680.408993] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.408993] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] return self._exit_event.wait() [ 680.408993] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 680.408993] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] current.throw(*self._exc) [ 680.409395] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.409395] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] result = function(*args, **kwargs) [ 680.409395] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.409395] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] return func(*args, **kwargs) [ 680.409395] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.409395] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] raise e [ 680.409395] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.409395] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] nwinfo = self.network_api.allocate_for_instance( [ 680.409395] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.409395] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] created_port_ids = self._update_ports_for_instance( [ 680.409395] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.409395] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] with excutils.save_and_reraise_exception(): [ 680.409395] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.409694] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] self.force_reraise() [ 680.409694] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.409694] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] raise self.value [ 680.409694] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.409694] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] updated_port = self._update_port( [ 680.409694] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.409694] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] _ensure_no_port_binding_failure(port) [ 680.409694] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.409694] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] raise exception.PortBindingFailed(port_id=port['id']) [ 680.409694] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] nova.exception.PortBindingFailed: Binding failed for port cfd31541-18d4-4711-80f4-ee76d04de9e9, please check neutron logs for more information. [ 680.409694] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] [ 680.409694] env[62600]: INFO nova.compute.manager [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Terminating instance [ 680.410866] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Acquiring lock "refresh_cache-932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.411035] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Acquired lock "refresh_cache-932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.411208] env[62600]: DEBUG nova.network.neutron [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 680.540323] env[62600]: DEBUG oslo_concurrency.lockutils [req-2de8d5c5-ebed-4820-902e-5ff24de6b27d req-3d595eb9-7202-4275-816e-2dfa3f45e41e service nova] Releasing lock "refresh_cache-3d19d349-e450-4307-a763-6269dcb06544" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.540569] env[62600]: DEBUG nova.compute.manager [req-2de8d5c5-ebed-4820-902e-5ff24de6b27d req-3d595eb9-7202-4275-816e-2dfa3f45e41e service nova] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Received event network-vif-deleted-15dceed5-8643-4f6b-ba3c-0ec3b34f0684 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 680.864766] env[62600]: DEBUG oslo_concurrency.lockutils [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.544s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.865325] env[62600]: DEBUG nova.compute.manager [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 680.870151] env[62600]: DEBUG oslo_concurrency.lockutils [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.390s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.936338] env[62600]: DEBUG nova.network.neutron [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.005145] env[62600]: DEBUG nova.network.neutron [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.373656] env[62600]: DEBUG nova.compute.utils [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 681.379406] env[62600]: DEBUG nova.compute.manager [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 681.379406] env[62600]: DEBUG nova.network.neutron [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 681.421054] env[62600]: DEBUG nova.policy [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f67e3448814a4ce58be9db14c8d99fe5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a83665bc260f46988ac33d82e961c2b8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 681.443684] env[62600]: DEBUG nova.compute.manager [req-5e2b70e5-b87b-4121-afdc-a2b7058c0262 req-8e1209d7-a1c8-4400-a7ff-5f75ec305505 service nova] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Received event network-changed-cfd31541-18d4-4711-80f4-ee76d04de9e9 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 681.443966] env[62600]: DEBUG nova.compute.manager [req-5e2b70e5-b87b-4121-afdc-a2b7058c0262 req-8e1209d7-a1c8-4400-a7ff-5f75ec305505 service nova] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Refreshing instance network info cache due to event network-changed-cfd31541-18d4-4711-80f4-ee76d04de9e9. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 681.444223] env[62600]: DEBUG oslo_concurrency.lockutils [req-5e2b70e5-b87b-4121-afdc-a2b7058c0262 req-8e1209d7-a1c8-4400-a7ff-5f75ec305505 service nova] Acquiring lock "refresh_cache-932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.509262] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Releasing lock "refresh_cache-932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.509674] env[62600]: DEBUG nova.compute.manager [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 681.510275] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 681.510376] env[62600]: DEBUG oslo_concurrency.lockutils [req-5e2b70e5-b87b-4121-afdc-a2b7058c0262 req-8e1209d7-a1c8-4400-a7ff-5f75ec305505 service nova] Acquired lock "refresh_cache-932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.510561] env[62600]: DEBUG nova.network.neutron [req-5e2b70e5-b87b-4121-afdc-a2b7058c0262 req-8e1209d7-a1c8-4400-a7ff-5f75ec305505 service nova] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Refreshing network info cache for port cfd31541-18d4-4711-80f4-ee76d04de9e9 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 681.511745] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a99736da-f8d5-4145-87b6-c552ea578a16 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.520819] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58898eaf-bf33-4dee-8708-5997b9bcb879 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.548319] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6 could not be found. [ 681.548539] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 681.548720] env[62600]: INFO nova.compute.manager [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 681.548956] env[62600]: DEBUG oslo.service.loopingcall [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 681.551527] env[62600]: DEBUG nova.compute.manager [-] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 681.551614] env[62600]: DEBUG nova.network.neutron [-] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 681.575544] env[62600]: DEBUG nova.network.neutron [-] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.758822] env[62600]: DEBUG nova.network.neutron [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Successfully created port: 776c7599-bde8-4d84-9c07-5796cc622714 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 681.801354] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef922fd3-4102-4613-a454-ed2eea3a83ba {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.808987] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8356aa02-9ae9-46b7-b9bf-538fcf94642a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.839065] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21237d9a-31e0-480d-843a-dd4a738d7cd2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.846655] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf5abaa-9597-49dc-bcc5-c50040d64e58 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.861248] env[62600]: DEBUG nova.compute.provider_tree [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.881091] env[62600]: DEBUG nova.compute.manager [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 682.040493] env[62600]: DEBUG nova.network.neutron [req-5e2b70e5-b87b-4121-afdc-a2b7058c0262 req-8e1209d7-a1c8-4400-a7ff-5f75ec305505 service nova] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 682.080275] env[62600]: DEBUG nova.network.neutron [-] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.215726] env[62600]: DEBUG nova.network.neutron [req-5e2b70e5-b87b-4121-afdc-a2b7058c0262 req-8e1209d7-a1c8-4400-a7ff-5f75ec305505 service nova] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.363773] env[62600]: DEBUG nova.scheduler.client.report [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 682.584050] env[62600]: INFO nova.compute.manager [-] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Took 1.03 seconds to deallocate network for instance. [ 682.587665] env[62600]: DEBUG nova.compute.claims [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 682.587873] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.721367] env[62600]: DEBUG oslo_concurrency.lockutils [req-5e2b70e5-b87b-4121-afdc-a2b7058c0262 req-8e1209d7-a1c8-4400-a7ff-5f75ec305505 service nova] Releasing lock "refresh_cache-932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.721696] env[62600]: DEBUG nova.compute.manager [req-5e2b70e5-b87b-4121-afdc-a2b7058c0262 req-8e1209d7-a1c8-4400-a7ff-5f75ec305505 service nova] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Received event network-vif-deleted-cfd31541-18d4-4711-80f4-ee76d04de9e9 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 682.868325] env[62600]: DEBUG oslo_concurrency.lockutils [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.999s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.868975] env[62600]: ERROR nova.compute.manager [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e4f492e1-f18e-4fbd-8c1f-bab956693c6d, please check neutron logs for more information. [ 682.868975] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] Traceback (most recent call last): [ 682.868975] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 682.868975] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] self.driver.spawn(context, instance, image_meta, [ 682.868975] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 682.868975] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] self._vmops.spawn(context, instance, image_meta, injected_files, [ 682.868975] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 682.868975] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] vm_ref = self.build_virtual_machine(instance, [ 682.868975] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 682.868975] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] vif_infos = vmwarevif.get_vif_info(self._session, [ 682.868975] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 682.869273] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] for vif in network_info: [ 682.869273] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 682.869273] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] return self._sync_wrapper(fn, *args, **kwargs) [ 682.869273] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 682.869273] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] self.wait() [ 682.869273] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 682.869273] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] self[:] = self._gt.wait() [ 682.869273] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 682.869273] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] return self._exit_event.wait() [ 682.869273] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 682.869273] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] result = hub.switch() [ 682.869273] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 682.869273] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] return self.greenlet.switch() [ 682.869569] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.869569] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] result = function(*args, **kwargs) [ 682.869569] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 682.869569] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] return func(*args, **kwargs) [ 682.869569] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 682.869569] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] raise e [ 682.869569] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.869569] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] nwinfo = self.network_api.allocate_for_instance( [ 682.869569] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 682.869569] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] created_port_ids = self._update_ports_for_instance( [ 682.869569] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 682.869569] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] with excutils.save_and_reraise_exception(): [ 682.869569] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.869872] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] self.force_reraise() [ 682.869872] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.869872] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] raise self.value [ 682.869872] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 682.869872] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] updated_port = self._update_port( [ 682.869872] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.869872] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] _ensure_no_port_binding_failure(port) [ 682.869872] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.869872] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] raise exception.PortBindingFailed(port_id=port['id']) [ 682.869872] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] nova.exception.PortBindingFailed: Binding failed for port e4f492e1-f18e-4fbd-8c1f-bab956693c6d, please check neutron logs for more information. [ 682.869872] env[62600]: ERROR nova.compute.manager [instance: 4455c247-a707-4c3a-9323-0110a1290780] [ 682.870235] env[62600]: DEBUG nova.compute.utils [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Binding failed for port e4f492e1-f18e-4fbd-8c1f-bab956693c6d, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 682.871427] env[62600]: DEBUG oslo_concurrency.lockutils [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.841s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.874073] env[62600]: DEBUG nova.compute.manager [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Build of instance 4455c247-a707-4c3a-9323-0110a1290780 was re-scheduled: Binding failed for port e4f492e1-f18e-4fbd-8c1f-bab956693c6d, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 682.874528] env[62600]: DEBUG nova.compute.manager [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 682.874752] env[62600]: DEBUG oslo_concurrency.lockutils [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Acquiring lock "refresh_cache-4455c247-a707-4c3a-9323-0110a1290780" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.874899] env[62600]: DEBUG oslo_concurrency.lockutils [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Acquired lock "refresh_cache-4455c247-a707-4c3a-9323-0110a1290780" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.875070] env[62600]: DEBUG nova.network.neutron [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 682.890907] env[62600]: DEBUG nova.compute.manager [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 682.916109] env[62600]: DEBUG nova.virt.hardware [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 682.916328] env[62600]: DEBUG nova.virt.hardware [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 682.916451] env[62600]: DEBUG nova.virt.hardware [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 682.916634] env[62600]: DEBUG nova.virt.hardware [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 682.916776] env[62600]: DEBUG nova.virt.hardware [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 682.916921] env[62600]: DEBUG nova.virt.hardware [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 682.917168] env[62600]: DEBUG nova.virt.hardware [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 682.917356] env[62600]: DEBUG nova.virt.hardware [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 682.917529] env[62600]: DEBUG nova.virt.hardware [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 682.917691] env[62600]: DEBUG nova.virt.hardware [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 682.917859] env[62600]: DEBUG nova.virt.hardware [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 682.918938] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13efe6c7-de5a-4771-9b29-10f12d885e16 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.929813] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-987439a2-c7a5-4eac-bc05-243f6f4e7605 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.406959] env[62600]: DEBUG nova.network.neutron [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 683.425511] env[62600]: ERROR nova.compute.manager [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 776c7599-bde8-4d84-9c07-5796cc622714, please check neutron logs for more information. [ 683.425511] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 683.425511] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.425511] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 683.425511] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 683.425511] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 683.425511] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 683.425511] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 683.425511] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.425511] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 683.425511] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.425511] env[62600]: ERROR nova.compute.manager raise self.value [ 683.425511] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 683.425511] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 683.425511] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.425511] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 683.426091] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.426091] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 683.426091] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 776c7599-bde8-4d84-9c07-5796cc622714, please check neutron logs for more information. [ 683.426091] env[62600]: ERROR nova.compute.manager [ 683.426091] env[62600]: Traceback (most recent call last): [ 683.426091] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 683.426091] env[62600]: listener.cb(fileno) [ 683.426091] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 683.426091] env[62600]: result = function(*args, **kwargs) [ 683.426091] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 683.426091] env[62600]: return func(*args, **kwargs) [ 683.426091] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 683.426091] env[62600]: raise e [ 683.426091] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.426091] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 683.426091] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 683.426091] env[62600]: created_port_ids = self._update_ports_for_instance( [ 683.426091] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 683.426091] env[62600]: with excutils.save_and_reraise_exception(): [ 683.426091] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.426091] env[62600]: self.force_reraise() [ 683.426091] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.426091] env[62600]: raise self.value [ 683.426091] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 683.426091] env[62600]: updated_port = self._update_port( [ 683.426091] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.426091] env[62600]: _ensure_no_port_binding_failure(port) [ 683.426091] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.426091] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 683.426904] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 776c7599-bde8-4d84-9c07-5796cc622714, please check neutron logs for more information. [ 683.426904] env[62600]: Removing descriptor: 16 [ 683.426904] env[62600]: ERROR nova.compute.manager [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 776c7599-bde8-4d84-9c07-5796cc622714, please check neutron logs for more information. [ 683.426904] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Traceback (most recent call last): [ 683.426904] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 683.426904] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] yield resources [ 683.426904] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 683.426904] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] self.driver.spawn(context, instance, image_meta, [ 683.426904] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 683.426904] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 683.426904] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 683.426904] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] vm_ref = self.build_virtual_machine(instance, [ 683.427287] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 683.427287] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 683.427287] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 683.427287] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] for vif in network_info: [ 683.427287] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 683.427287] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] return self._sync_wrapper(fn, *args, **kwargs) [ 683.427287] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 683.427287] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] self.wait() [ 683.427287] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 683.427287] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] self[:] = self._gt.wait() [ 683.427287] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 683.427287] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] return self._exit_event.wait() [ 683.427287] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 683.427644] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] result = hub.switch() [ 683.427644] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 683.427644] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] return self.greenlet.switch() [ 683.427644] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 683.427644] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] result = function(*args, **kwargs) [ 683.427644] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 683.427644] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] return func(*args, **kwargs) [ 683.427644] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 683.427644] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] raise e [ 683.427644] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.427644] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] nwinfo = self.network_api.allocate_for_instance( [ 683.427644] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 683.427644] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] created_port_ids = self._update_ports_for_instance( [ 683.427979] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 683.427979] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] with excutils.save_and_reraise_exception(): [ 683.427979] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.427979] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] self.force_reraise() [ 683.427979] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.427979] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] raise self.value [ 683.427979] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 683.427979] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] updated_port = self._update_port( [ 683.427979] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.427979] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] _ensure_no_port_binding_failure(port) [ 683.427979] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.427979] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] raise exception.PortBindingFailed(port_id=port['id']) [ 683.428391] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] nova.exception.PortBindingFailed: Binding failed for port 776c7599-bde8-4d84-9c07-5796cc622714, please check neutron logs for more information. [ 683.428391] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] [ 683.428391] env[62600]: INFO nova.compute.manager [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Terminating instance [ 683.428391] env[62600]: DEBUG oslo_concurrency.lockutils [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Acquiring lock "refresh_cache-20fab902-79cc-4f83-abaf-53a53c2521f4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.428514] env[62600]: DEBUG oslo_concurrency.lockutils [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Acquired lock "refresh_cache-20fab902-79cc-4f83-abaf-53a53c2521f4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.430545] env[62600]: DEBUG nova.network.neutron [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 683.474454] env[62600]: DEBUG nova.compute.manager [req-27e5c55f-6bb2-4c54-adcc-863e32fa10de req-7670d2e4-f885-48a6-bfa8-5b145371c2f0 service nova] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Received event network-changed-776c7599-bde8-4d84-9c07-5796cc622714 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 683.474645] env[62600]: DEBUG nova.compute.manager [req-27e5c55f-6bb2-4c54-adcc-863e32fa10de req-7670d2e4-f885-48a6-bfa8-5b145371c2f0 service nova] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Refreshing instance network info cache due to event network-changed-776c7599-bde8-4d84-9c07-5796cc622714. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 683.475112] env[62600]: DEBUG oslo_concurrency.lockutils [req-27e5c55f-6bb2-4c54-adcc-863e32fa10de req-7670d2e4-f885-48a6-bfa8-5b145371c2f0 service nova] Acquiring lock "refresh_cache-20fab902-79cc-4f83-abaf-53a53c2521f4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.537690] env[62600]: DEBUG nova.network.neutron [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.799967] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea4a943-3eae-4150-a63e-6df29e61ef5b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.808272] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed04bf7-1ed3-4752-8c0c-129c0da93efa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.840336] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b1a1563-dcd8-47a3-a9e7-374a60298ae6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.848548] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76fde118-d154-4e7f-84fe-1b10eeb4de13 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.861297] env[62600]: DEBUG nova.compute.provider_tree [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.951031] env[62600]: DEBUG nova.network.neutron [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 684.040195] env[62600]: DEBUG oslo_concurrency.lockutils [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Releasing lock "refresh_cache-4455c247-a707-4c3a-9323-0110a1290780" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.040508] env[62600]: DEBUG nova.compute.manager [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 684.040716] env[62600]: DEBUG nova.compute.manager [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 684.040896] env[62600]: DEBUG nova.network.neutron [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 684.046664] env[62600]: DEBUG nova.network.neutron [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.057629] env[62600]: DEBUG nova.network.neutron [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 684.367113] env[62600]: DEBUG nova.scheduler.client.report [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 684.549772] env[62600]: DEBUG oslo_concurrency.lockutils [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Releasing lock "refresh_cache-20fab902-79cc-4f83-abaf-53a53c2521f4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.550586] env[62600]: DEBUG nova.compute.manager [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 684.550586] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 684.550908] env[62600]: DEBUG oslo_concurrency.lockutils [req-27e5c55f-6bb2-4c54-adcc-863e32fa10de req-7670d2e4-f885-48a6-bfa8-5b145371c2f0 service nova] Acquired lock "refresh_cache-20fab902-79cc-4f83-abaf-53a53c2521f4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.551100] env[62600]: DEBUG nova.network.neutron [req-27e5c55f-6bb2-4c54-adcc-863e32fa10de req-7670d2e4-f885-48a6-bfa8-5b145371c2f0 service nova] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Refreshing network info cache for port 776c7599-bde8-4d84-9c07-5796cc622714 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 684.552530] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dc0f9f7b-43a3-45d0-afab-7960eca80102 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.562483] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0f71dce-9d9d-41f2-8a97-7e6a8b1c766e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.573404] env[62600]: DEBUG nova.network.neutron [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.587760] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 20fab902-79cc-4f83-abaf-53a53c2521f4 could not be found. [ 684.587871] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 684.588069] env[62600]: INFO nova.compute.manager [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 684.588343] env[62600]: DEBUG oslo.service.loopingcall [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 684.589115] env[62600]: DEBUG nova.compute.manager [-] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 684.589303] env[62600]: DEBUG nova.network.neutron [-] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 684.606952] env[62600]: DEBUG nova.network.neutron [-] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 684.874707] env[62600]: DEBUG oslo_concurrency.lockutils [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.003s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.875459] env[62600]: ERROR nova.compute.manager [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 556c03da-1c62-48dc-a7f4-d107804f7a59, please check neutron logs for more information. [ 684.875459] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Traceback (most recent call last): [ 684.875459] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 684.875459] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] self.driver.spawn(context, instance, image_meta, [ 684.875459] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 684.875459] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.875459] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.875459] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] vm_ref = self.build_virtual_machine(instance, [ 684.875459] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.875459] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.875459] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.875745] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] for vif in network_info: [ 684.875745] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 684.875745] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] return self._sync_wrapper(fn, *args, **kwargs) [ 684.875745] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 684.875745] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] self.wait() [ 684.875745] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 684.875745] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] self[:] = self._gt.wait() [ 684.875745] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.875745] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] return self._exit_event.wait() [ 684.875745] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 684.875745] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] result = hub.switch() [ 684.875745] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 684.875745] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] return self.greenlet.switch() [ 684.876499] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.876499] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] result = function(*args, **kwargs) [ 684.876499] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 684.876499] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] return func(*args, **kwargs) [ 684.876499] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.876499] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] raise e [ 684.876499] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.876499] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] nwinfo = self.network_api.allocate_for_instance( [ 684.876499] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 684.876499] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] created_port_ids = self._update_ports_for_instance( [ 684.876499] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 684.876499] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] with excutils.save_and_reraise_exception(): [ 684.876499] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.876933] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] self.force_reraise() [ 684.876933] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.876933] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] raise self.value [ 684.876933] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 684.876933] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] updated_port = self._update_port( [ 684.876933] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.876933] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] _ensure_no_port_binding_failure(port) [ 684.876933] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.876933] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] raise exception.PortBindingFailed(port_id=port['id']) [ 684.876933] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] nova.exception.PortBindingFailed: Binding failed for port 556c03da-1c62-48dc-a7f4-d107804f7a59, please check neutron logs for more information. [ 684.876933] env[62600]: ERROR nova.compute.manager [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] [ 684.877209] env[62600]: DEBUG nova.compute.utils [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Binding failed for port 556c03da-1c62-48dc-a7f4-d107804f7a59, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 684.877643] env[62600]: DEBUG oslo_concurrency.lockutils [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.357s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.879398] env[62600]: INFO nova.compute.claims [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 684.882557] env[62600]: DEBUG nova.compute.manager [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Build of instance 09ccf23b-4763-47e1-afad-59e2b06b2d9e was re-scheduled: Binding failed for port 556c03da-1c62-48dc-a7f4-d107804f7a59, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 684.882977] env[62600]: DEBUG nova.compute.manager [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 684.883218] env[62600]: DEBUG oslo_concurrency.lockutils [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Acquiring lock "refresh_cache-09ccf23b-4763-47e1-afad-59e2b06b2d9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.883369] env[62600]: DEBUG oslo_concurrency.lockutils [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Acquired lock "refresh_cache-09ccf23b-4763-47e1-afad-59e2b06b2d9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.883531] env[62600]: DEBUG nova.network.neutron [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 685.073434] env[62600]: DEBUG nova.network.neutron [req-27e5c55f-6bb2-4c54-adcc-863e32fa10de req-7670d2e4-f885-48a6-bfa8-5b145371c2f0 service nova] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 685.075645] env[62600]: INFO nova.compute.manager [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] [instance: 4455c247-a707-4c3a-9323-0110a1290780] Took 1.03 seconds to deallocate network for instance. [ 685.110849] env[62600]: DEBUG nova.network.neutron [-] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.157393] env[62600]: DEBUG nova.network.neutron [req-27e5c55f-6bb2-4c54-adcc-863e32fa10de req-7670d2e4-f885-48a6-bfa8-5b145371c2f0 service nova] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.403567] env[62600]: DEBUG nova.network.neutron [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 685.448273] env[62600]: DEBUG nova.network.neutron [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.613625] env[62600]: INFO nova.compute.manager [-] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Took 1.02 seconds to deallocate network for instance. [ 685.615897] env[62600]: DEBUG nova.compute.claims [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 685.616154] env[62600]: DEBUG oslo_concurrency.lockutils [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.663064] env[62600]: DEBUG oslo_concurrency.lockutils [req-27e5c55f-6bb2-4c54-adcc-863e32fa10de req-7670d2e4-f885-48a6-bfa8-5b145371c2f0 service nova] Releasing lock "refresh_cache-20fab902-79cc-4f83-abaf-53a53c2521f4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.663402] env[62600]: DEBUG nova.compute.manager [req-27e5c55f-6bb2-4c54-adcc-863e32fa10de req-7670d2e4-f885-48a6-bfa8-5b145371c2f0 service nova] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Received event network-vif-deleted-776c7599-bde8-4d84-9c07-5796cc622714 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 685.951047] env[62600]: DEBUG oslo_concurrency.lockutils [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Releasing lock "refresh_cache-09ccf23b-4763-47e1-afad-59e2b06b2d9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.951047] env[62600]: DEBUG nova.compute.manager [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 685.951047] env[62600]: DEBUG nova.compute.manager [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 685.951047] env[62600]: DEBUG nova.network.neutron [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 685.966908] env[62600]: DEBUG nova.network.neutron [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 686.101616] env[62600]: INFO nova.scheduler.client.report [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Deleted allocations for instance 4455c247-a707-4c3a-9323-0110a1290780 [ 686.231304] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20def984-b43c-4c0b-9c47-295869159044 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.238722] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd0fba03-2f36-46eb-aae2-74c2d6e7646e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.269130] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c462afdc-20a8-49d3-81a0-d6cffab338f2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.276581] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aced416b-7c91-4917-80d3-d51775d46943 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.290857] env[62600]: DEBUG nova.compute.provider_tree [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.471581] env[62600]: DEBUG nova.network.neutron [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.612658] env[62600]: DEBUG oslo_concurrency.lockutils [None req-382b6f69-04a2-44f3-95b0-e495492e8536 tempest-FloatingIPsAssociationNegativeTestJSON-160074039 tempest-FloatingIPsAssociationNegativeTestJSON-160074039-project-member] Lock "4455c247-a707-4c3a-9323-0110a1290780" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.037s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.794462] env[62600]: DEBUG nova.scheduler.client.report [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 686.974550] env[62600]: INFO nova.compute.manager [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] [instance: 09ccf23b-4763-47e1-afad-59e2b06b2d9e] Took 1.02 seconds to deallocate network for instance. [ 687.116501] env[62600]: DEBUG nova.compute.manager [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 687.299244] env[62600]: DEBUG oslo_concurrency.lockutils [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.299815] env[62600]: DEBUG nova.compute.manager [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 687.302536] env[62600]: DEBUG oslo_concurrency.lockutils [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.815s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.636248] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.806935] env[62600]: DEBUG nova.compute.utils [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 687.811283] env[62600]: DEBUG nova.compute.manager [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 687.811454] env[62600]: DEBUG nova.network.neutron [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 687.853555] env[62600]: DEBUG nova.policy [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cb6b500f66c249629e0bb20838dca631', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '88985cba32da4faaa133d74ffa635292', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 688.008920] env[62600]: INFO nova.scheduler.client.report [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Deleted allocations for instance 09ccf23b-4763-47e1-afad-59e2b06b2d9e [ 688.190336] env[62600]: DEBUG nova.network.neutron [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Successfully created port: e97d0860-1801-4d1e-b455-62b34fa87099 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 688.222119] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62267706-160f-44ef-9311-511667ca979a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.229850] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff06e8aa-041e-4433-9b4a-49f5fd73064b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.260414] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee7cc01f-3f43-45a9-b995-656beefa295c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.267747] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc634c1-0b64-4ace-8223-b38f7c84ed6d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.281244] env[62600]: DEBUG nova.compute.provider_tree [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 688.311885] env[62600]: DEBUG nova.compute.manager [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 688.517456] env[62600]: DEBUG oslo_concurrency.lockutils [None req-78bfdf68-dfcf-49aa-bd9e-9509e87a11f4 tempest-ServerDiagnosticsNegativeTest-1637034673 tempest-ServerDiagnosticsNegativeTest-1637034673-project-member] Lock "09ccf23b-4763-47e1-afad-59e2b06b2d9e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.871s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.788215] env[62600]: DEBUG nova.scheduler.client.report [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 689.021702] env[62600]: DEBUG nova.compute.manager [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 689.270167] env[62600]: DEBUG nova.compute.manager [req-f83b93e7-1468-4f40-ab37-954802c02974 req-d6005dfd-0268-410d-a493-f0b257798ddb service nova] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Received event network-changed-e97d0860-1801-4d1e-b455-62b34fa87099 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 689.270379] env[62600]: DEBUG nova.compute.manager [req-f83b93e7-1468-4f40-ab37-954802c02974 req-d6005dfd-0268-410d-a493-f0b257798ddb service nova] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Refreshing instance network info cache due to event network-changed-e97d0860-1801-4d1e-b455-62b34fa87099. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 689.270557] env[62600]: DEBUG oslo_concurrency.lockutils [req-f83b93e7-1468-4f40-ab37-954802c02974 req-d6005dfd-0268-410d-a493-f0b257798ddb service nova] Acquiring lock "refresh_cache-6b19c327-9a80-498f-a350-8068ad250377" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.270722] env[62600]: DEBUG oslo_concurrency.lockutils [req-f83b93e7-1468-4f40-ab37-954802c02974 req-d6005dfd-0268-410d-a493-f0b257798ddb service nova] Acquired lock "refresh_cache-6b19c327-9a80-498f-a350-8068ad250377" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.270848] env[62600]: DEBUG nova.network.neutron [req-f83b93e7-1468-4f40-ab37-954802c02974 req-d6005dfd-0268-410d-a493-f0b257798ddb service nova] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Refreshing network info cache for port e97d0860-1801-4d1e-b455-62b34fa87099 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 689.286700] env[62600]: ERROR nova.compute.manager [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e97d0860-1801-4d1e-b455-62b34fa87099, please check neutron logs for more information. [ 689.286700] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 689.286700] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.286700] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 689.286700] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 689.286700] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 689.286700] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 689.286700] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 689.286700] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.286700] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 689.286700] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.286700] env[62600]: ERROR nova.compute.manager raise self.value [ 689.286700] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 689.286700] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 689.286700] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.286700] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 689.287500] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.287500] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 689.287500] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e97d0860-1801-4d1e-b455-62b34fa87099, please check neutron logs for more information. [ 689.287500] env[62600]: ERROR nova.compute.manager [ 689.287500] env[62600]: Traceback (most recent call last): [ 689.287500] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 689.287500] env[62600]: listener.cb(fileno) [ 689.287500] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.287500] env[62600]: result = function(*args, **kwargs) [ 689.287500] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.287500] env[62600]: return func(*args, **kwargs) [ 689.287500] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.287500] env[62600]: raise e [ 689.287500] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.287500] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 689.287500] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 689.287500] env[62600]: created_port_ids = self._update_ports_for_instance( [ 689.287500] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 689.287500] env[62600]: with excutils.save_and_reraise_exception(): [ 689.287500] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.287500] env[62600]: self.force_reraise() [ 689.287500] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.287500] env[62600]: raise self.value [ 689.287500] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 689.287500] env[62600]: updated_port = self._update_port( [ 689.287500] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.287500] env[62600]: _ensure_no_port_binding_failure(port) [ 689.287500] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.287500] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 689.288763] env[62600]: nova.exception.PortBindingFailed: Binding failed for port e97d0860-1801-4d1e-b455-62b34fa87099, please check neutron logs for more information. [ 689.288763] env[62600]: Removing descriptor: 16 [ 689.291610] env[62600]: DEBUG oslo_concurrency.lockutils [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.989s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.292266] env[62600]: ERROR nova.compute.manager [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c13d7fb7-a219-4c26-8958-fc34681c7277, please check neutron logs for more information. [ 689.292266] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] Traceback (most recent call last): [ 689.292266] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 689.292266] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] self.driver.spawn(context, instance, image_meta, [ 689.292266] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 689.292266] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 689.292266] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 689.292266] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] vm_ref = self.build_virtual_machine(instance, [ 689.292266] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 689.292266] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] vif_infos = vmwarevif.get_vif_info(self._session, [ 689.292266] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 689.292681] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] for vif in network_info: [ 689.292681] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 689.292681] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] return self._sync_wrapper(fn, *args, **kwargs) [ 689.292681] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 689.292681] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] self.wait() [ 689.292681] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 689.292681] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] self[:] = self._gt.wait() [ 689.292681] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 689.292681] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] return self._exit_event.wait() [ 689.292681] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 689.292681] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] current.throw(*self._exc) [ 689.292681] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.292681] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] result = function(*args, **kwargs) [ 689.293104] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.293104] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] return func(*args, **kwargs) [ 689.293104] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.293104] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] raise e [ 689.293104] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.293104] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] nwinfo = self.network_api.allocate_for_instance( [ 689.293104] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 689.293104] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] created_port_ids = self._update_ports_for_instance( [ 689.293104] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 689.293104] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] with excutils.save_and_reraise_exception(): [ 689.293104] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.293104] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] self.force_reraise() [ 689.293104] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.293452] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] raise self.value [ 689.293452] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 689.293452] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] updated_port = self._update_port( [ 689.293452] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.293452] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] _ensure_no_port_binding_failure(port) [ 689.293452] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.293452] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] raise exception.PortBindingFailed(port_id=port['id']) [ 689.293452] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] nova.exception.PortBindingFailed: Binding failed for port c13d7fb7-a219-4c26-8958-fc34681c7277, please check neutron logs for more information. [ 689.293452] env[62600]: ERROR nova.compute.manager [instance: 60db0e43-692b-4449-9b44-badf3053429d] [ 689.293452] env[62600]: DEBUG nova.compute.utils [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Binding failed for port c13d7fb7-a219-4c26-8958-fc34681c7277, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 689.295350] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.067s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.300954] env[62600]: DEBUG nova.compute.manager [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Build of instance 60db0e43-692b-4449-9b44-badf3053429d was re-scheduled: Binding failed for port c13d7fb7-a219-4c26-8958-fc34681c7277, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 689.300954] env[62600]: DEBUG nova.compute.manager [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 689.300954] env[62600]: DEBUG oslo_concurrency.lockutils [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Acquiring lock "refresh_cache-60db0e43-692b-4449-9b44-badf3053429d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.300954] env[62600]: DEBUG oslo_concurrency.lockutils [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Acquired lock "refresh_cache-60db0e43-692b-4449-9b44-badf3053429d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.301130] env[62600]: DEBUG nova.network.neutron [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 689.320413] env[62600]: DEBUG nova.compute.manager [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 689.353131] env[62600]: DEBUG nova.virt.hardware [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 689.353131] env[62600]: DEBUG nova.virt.hardware [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 689.353423] env[62600]: DEBUG nova.virt.hardware [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 689.353458] env[62600]: DEBUG nova.virt.hardware [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 689.354082] env[62600]: DEBUG nova.virt.hardware [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 689.354082] env[62600]: DEBUG nova.virt.hardware [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 689.354082] env[62600]: DEBUG nova.virt.hardware [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 689.354251] env[62600]: DEBUG nova.virt.hardware [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 689.354541] env[62600]: DEBUG nova.virt.hardware [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 689.354541] env[62600]: DEBUG nova.virt.hardware [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 689.354641] env[62600]: DEBUG nova.virt.hardware [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 689.355689] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a4229f-7ace-4251-9eef-0180346f3c5f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.363749] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca5fc3b-e25c-49e7-acb2-c496b7a326a0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.377773] env[62600]: ERROR nova.compute.manager [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e97d0860-1801-4d1e-b455-62b34fa87099, please check neutron logs for more information. [ 689.377773] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] Traceback (most recent call last): [ 689.377773] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 689.377773] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] yield resources [ 689.377773] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 689.377773] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] self.driver.spawn(context, instance, image_meta, [ 689.377773] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 689.377773] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] self._vmops.spawn(context, instance, image_meta, injected_files, [ 689.377773] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 689.377773] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] vm_ref = self.build_virtual_machine(instance, [ 689.377773] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 689.378140] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] vif_infos = vmwarevif.get_vif_info(self._session, [ 689.378140] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 689.378140] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] for vif in network_info: [ 689.378140] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 689.378140] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] return self._sync_wrapper(fn, *args, **kwargs) [ 689.378140] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 689.378140] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] self.wait() [ 689.378140] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 689.378140] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] self[:] = self._gt.wait() [ 689.378140] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 689.378140] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] return self._exit_event.wait() [ 689.378140] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 689.378140] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] current.throw(*self._exc) [ 689.378491] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.378491] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] result = function(*args, **kwargs) [ 689.378491] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.378491] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] return func(*args, **kwargs) [ 689.378491] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.378491] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] raise e [ 689.378491] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.378491] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] nwinfo = self.network_api.allocate_for_instance( [ 689.378491] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 689.378491] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] created_port_ids = self._update_ports_for_instance( [ 689.378491] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 689.378491] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] with excutils.save_and_reraise_exception(): [ 689.378491] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.378872] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] self.force_reraise() [ 689.378872] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.378872] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] raise self.value [ 689.378872] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 689.378872] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] updated_port = self._update_port( [ 689.378872] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.378872] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] _ensure_no_port_binding_failure(port) [ 689.378872] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.378872] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] raise exception.PortBindingFailed(port_id=port['id']) [ 689.378872] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] nova.exception.PortBindingFailed: Binding failed for port e97d0860-1801-4d1e-b455-62b34fa87099, please check neutron logs for more information. [ 689.378872] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] [ 689.378872] env[62600]: INFO nova.compute.manager [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Terminating instance [ 689.380166] env[62600]: DEBUG oslo_concurrency.lockutils [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquiring lock "refresh_cache-6b19c327-9a80-498f-a350-8068ad250377" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.550841] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.791922] env[62600]: DEBUG nova.network.neutron [req-f83b93e7-1468-4f40-ab37-954802c02974 req-d6005dfd-0268-410d-a493-f0b257798ddb service nova] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 689.822890] env[62600]: DEBUG nova.network.neutron [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 689.973792] env[62600]: DEBUG nova.network.neutron [req-f83b93e7-1468-4f40-ab37-954802c02974 req-d6005dfd-0268-410d-a493-f0b257798ddb service nova] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.979129] env[62600]: DEBUG nova.network.neutron [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.261424] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b20eaf39-f699-4914-a98f-3a7ce121e2d2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.269229] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87fc6a7f-f6a6-4eb2-ac54-ab30c025b40d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.304448] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd7ba610-4fcf-4b3a-a518-051928b7a7bc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.312567] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-702cf82f-1426-431f-a570-941b56fa9b6d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.327483] env[62600]: DEBUG nova.compute.provider_tree [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 690.488751] env[62600]: DEBUG oslo_concurrency.lockutils [req-f83b93e7-1468-4f40-ab37-954802c02974 req-d6005dfd-0268-410d-a493-f0b257798ddb service nova] Releasing lock "refresh_cache-6b19c327-9a80-498f-a350-8068ad250377" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.489346] env[62600]: DEBUG oslo_concurrency.lockutils [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Releasing lock "refresh_cache-60db0e43-692b-4449-9b44-badf3053429d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.489556] env[62600]: DEBUG nova.compute.manager [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 690.491842] env[62600]: DEBUG nova.compute.manager [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 690.491842] env[62600]: DEBUG nova.network.neutron [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 690.492565] env[62600]: DEBUG oslo_concurrency.lockutils [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquired lock "refresh_cache-6b19c327-9a80-498f-a350-8068ad250377" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.492782] env[62600]: DEBUG nova.network.neutron [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 690.508829] env[62600]: DEBUG nova.network.neutron [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 690.831190] env[62600]: DEBUG nova.scheduler.client.report [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 691.012112] env[62600]: DEBUG nova.network.neutron [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.014632] env[62600]: DEBUG nova.network.neutron [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 691.140100] env[62600]: DEBUG nova.network.neutron [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.310747] env[62600]: DEBUG nova.compute.manager [req-2ac1f9a2-eee9-4421-a929-9caa2187aa35 req-0cd11079-a89a-4da9-9d3e-899484586e58 service nova] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Received event network-vif-deleted-e97d0860-1801-4d1e-b455-62b34fa87099 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 691.342043] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.044s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.342043] env[62600]: ERROR nova.compute.manager [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c8f65221-5a81-4065-b966-78f0d85241ce, please check neutron logs for more information. [ 691.342043] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Traceback (most recent call last): [ 691.342043] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 691.342043] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] self.driver.spawn(context, instance, image_meta, [ 691.342043] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 691.342043] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 691.342043] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 691.342043] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] vm_ref = self.build_virtual_machine(instance, [ 691.342390] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 691.342390] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] vif_infos = vmwarevif.get_vif_info(self._session, [ 691.342390] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 691.342390] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] for vif in network_info: [ 691.342390] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 691.342390] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] return self._sync_wrapper(fn, *args, **kwargs) [ 691.342390] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 691.342390] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] self.wait() [ 691.342390] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 691.342390] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] self[:] = self._gt.wait() [ 691.342390] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 691.342390] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] return self._exit_event.wait() [ 691.342390] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 691.342732] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] current.throw(*self._exc) [ 691.342732] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.342732] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] result = function(*args, **kwargs) [ 691.342732] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 691.342732] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] return func(*args, **kwargs) [ 691.342732] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 691.342732] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] raise e [ 691.342732] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.342732] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] nwinfo = self.network_api.allocate_for_instance( [ 691.342732] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 691.342732] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] created_port_ids = self._update_ports_for_instance( [ 691.342732] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 691.342732] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] with excutils.save_and_reraise_exception(): [ 691.343108] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.343108] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] self.force_reraise() [ 691.343108] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.343108] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] raise self.value [ 691.343108] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 691.343108] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] updated_port = self._update_port( [ 691.343108] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.343108] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] _ensure_no_port_binding_failure(port) [ 691.343108] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.343108] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] raise exception.PortBindingFailed(port_id=port['id']) [ 691.343108] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] nova.exception.PortBindingFailed: Binding failed for port c8f65221-5a81-4065-b966-78f0d85241ce, please check neutron logs for more information. [ 691.343108] env[62600]: ERROR nova.compute.manager [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] [ 691.343471] env[62600]: DEBUG nova.compute.utils [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Binding failed for port c8f65221-5a81-4065-b966-78f0d85241ce, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 691.343787] env[62600]: DEBUG oslo_concurrency.lockutils [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.607s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.345356] env[62600]: INFO nova.compute.claims [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 691.349053] env[62600]: DEBUG nova.compute.manager [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Build of instance 9fd421c6-d068-4a81-a042-2d1f23fb9e1a was re-scheduled: Binding failed for port c8f65221-5a81-4065-b966-78f0d85241ce, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 691.349624] env[62600]: DEBUG nova.compute.manager [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 691.349945] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Acquiring lock "refresh_cache-9fd421c6-d068-4a81-a042-2d1f23fb9e1a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.350230] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Acquired lock "refresh_cache-9fd421c6-d068-4a81-a042-2d1f23fb9e1a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.350484] env[62600]: DEBUG nova.network.neutron [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 691.517130] env[62600]: INFO nova.compute.manager [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] [instance: 60db0e43-692b-4449-9b44-badf3053429d] Took 1.03 seconds to deallocate network for instance. [ 691.642468] env[62600]: DEBUG oslo_concurrency.lockutils [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Releasing lock "refresh_cache-6b19c327-9a80-498f-a350-8068ad250377" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.642937] env[62600]: DEBUG nova.compute.manager [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 691.643234] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 691.643479] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5e2fb062-ed78-4a61-93fd-a833ed02ed0a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.654335] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5f64dfe-f66a-4ef5-8db5-56e1f2c02b35 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.680818] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6b19c327-9a80-498f-a350-8068ad250377 could not be found. [ 691.681153] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 691.681296] env[62600]: INFO nova.compute.manager [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Took 0.04 seconds to destroy the instance on the hypervisor. [ 691.681822] env[62600]: DEBUG oslo.service.loopingcall [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 691.683016] env[62600]: DEBUG nova.compute.manager [-] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 691.683016] env[62600]: DEBUG nova.network.neutron [-] [instance: 6b19c327-9a80-498f-a350-8068ad250377] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 691.710464] env[62600]: DEBUG nova.network.neutron [-] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 691.875013] env[62600]: DEBUG nova.network.neutron [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 691.940247] env[62600]: DEBUG nova.network.neutron [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.218065] env[62600]: DEBUG nova.network.neutron [-] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.444791] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Releasing lock "refresh_cache-9fd421c6-d068-4a81-a042-2d1f23fb9e1a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.445172] env[62600]: DEBUG nova.compute.manager [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 692.445384] env[62600]: DEBUG nova.compute.manager [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 692.445642] env[62600]: DEBUG nova.network.neutron [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 692.465827] env[62600]: DEBUG nova.network.neutron [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 692.559560] env[62600]: INFO nova.scheduler.client.report [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Deleted allocations for instance 60db0e43-692b-4449-9b44-badf3053429d [ 692.720729] env[62600]: INFO nova.compute.manager [-] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Took 1.04 seconds to deallocate network for instance. [ 692.722982] env[62600]: DEBUG nova.compute.claims [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 692.723182] env[62600]: DEBUG oslo_concurrency.lockutils [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.856887] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eb4eeff-09e4-4f12-b6af-6cbe1e3f70c2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.865260] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d238842d-ae5b-4316-9c72-b7de5cf24952 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.897239] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fad0b9d-975c-42a2-8e21-b56348ac6afe {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.905096] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6b35ec6-7a3a-4c5d-b1ee-7b05ca332f58 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.918795] env[62600]: DEBUG nova.compute.provider_tree [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.969398] env[62600]: DEBUG nova.network.neutron [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.086184] env[62600]: DEBUG oslo_concurrency.lockutils [None req-729b9b44-61a8-4388-9224-e8b8230dc1ff tempest-InstanceActionsTestJSON-1083585951 tempest-InstanceActionsTestJSON-1083585951-project-member] Lock "60db0e43-692b-4449-9b44-badf3053429d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.117s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.422724] env[62600]: DEBUG nova.scheduler.client.report [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 693.472548] env[62600]: INFO nova.compute.manager [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] [instance: 9fd421c6-d068-4a81-a042-2d1f23fb9e1a] Took 1.03 seconds to deallocate network for instance. [ 693.584014] env[62600]: DEBUG nova.compute.manager [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 693.929153] env[62600]: DEBUG oslo_concurrency.lockutils [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.585s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.929691] env[62600]: DEBUG nova.compute.manager [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 693.932410] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.088s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.101935] env[62600]: DEBUG oslo_concurrency.lockutils [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.124772] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Acquiring lock "734e8a2b-faaa-4165-b0e9-69eeb3699b2d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.124873] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Lock "734e8a2b-faaa-4165-b0e9-69eeb3699b2d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.440555] env[62600]: DEBUG nova.compute.utils [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 694.446709] env[62600]: DEBUG nova.compute.manager [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 694.446939] env[62600]: DEBUG nova.network.neutron [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 694.499482] env[62600]: DEBUG nova.policy [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '213369dec97d4fac927ebd46f1201a64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b976c5b07a184e058b5b16530bdef27e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 694.516976] env[62600]: INFO nova.scheduler.client.report [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Deleted allocations for instance 9fd421c6-d068-4a81-a042-2d1f23fb9e1a [ 694.948212] env[62600]: DEBUG nova.compute.manager [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 694.986691] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance e8cbb437-0ace-4af3-a1d9-f96eeeea8504 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 694.986890] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 3d19d349-e450-4307-a763-6269dcb06544 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 694.987064] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 694.987197] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 20fab902-79cc-4f83-abaf-53a53c2521f4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 694.987341] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 6b19c327-9a80-498f-a350-8068ad250377 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 695.029208] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ad9a69be-38e5-47f9-8a76-8e72686184fd tempest-DeleteServersAdminTestJSON-626313264 tempest-DeleteServersAdminTestJSON-626313264-project-member] Lock "9fd421c6-d068-4a81-a042-2d1f23fb9e1a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.190s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.057097] env[62600]: DEBUG nova.network.neutron [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Successfully created port: d0ae7b46-9488-45b2-89ed-a66a603cdec1 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 695.491257] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance c93cb228-0adb-423c-854f-73354cdb7012 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 695.491421] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 695.532562] env[62600]: DEBUG nova.compute.manager [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 695.933451] env[62600]: DEBUG nova.compute.manager [req-4da20fbb-b23a-4151-83da-2dc812f6d29c req-39ea7f43-781c-4e18-ba13-23fee528d746 service nova] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Received event network-changed-d0ae7b46-9488-45b2-89ed-a66a603cdec1 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 695.933787] env[62600]: DEBUG nova.compute.manager [req-4da20fbb-b23a-4151-83da-2dc812f6d29c req-39ea7f43-781c-4e18-ba13-23fee528d746 service nova] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Refreshing instance network info cache due to event network-changed-d0ae7b46-9488-45b2-89ed-a66a603cdec1. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 695.933912] env[62600]: DEBUG oslo_concurrency.lockutils [req-4da20fbb-b23a-4151-83da-2dc812f6d29c req-39ea7f43-781c-4e18-ba13-23fee528d746 service nova] Acquiring lock "refresh_cache-15772d8e-98ce-47cf-9b61-e79c7ffc3a8b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.934017] env[62600]: DEBUG oslo_concurrency.lockutils [req-4da20fbb-b23a-4151-83da-2dc812f6d29c req-39ea7f43-781c-4e18-ba13-23fee528d746 service nova] Acquired lock "refresh_cache-15772d8e-98ce-47cf-9b61-e79c7ffc3a8b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.934197] env[62600]: DEBUG nova.network.neutron [req-4da20fbb-b23a-4151-83da-2dc812f6d29c req-39ea7f43-781c-4e18-ba13-23fee528d746 service nova] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Refreshing network info cache for port d0ae7b46-9488-45b2-89ed-a66a603cdec1 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 695.966611] env[62600]: DEBUG nova.compute.manager [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 695.998467] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance e3416076-1d11-47e4-9f76-9a64d7259166 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 696.006053] env[62600]: DEBUG nova.virt.hardware [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 696.006324] env[62600]: DEBUG nova.virt.hardware [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 696.006489] env[62600]: DEBUG nova.virt.hardware [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 696.006675] env[62600]: DEBUG nova.virt.hardware [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 696.006821] env[62600]: DEBUG nova.virt.hardware [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 696.006967] env[62600]: DEBUG nova.virt.hardware [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 696.007204] env[62600]: DEBUG nova.virt.hardware [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 696.007367] env[62600]: DEBUG nova.virt.hardware [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 696.007535] env[62600]: DEBUG nova.virt.hardware [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 696.007698] env[62600]: DEBUG nova.virt.hardware [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 696.007871] env[62600]: DEBUG nova.virt.hardware [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 696.009059] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81858ae1-ac83-4c33-bdd4-6937b3fdf481 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.017721] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e7faed-6249-4945-a472-55d4910ba65b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.064397] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.150862] env[62600]: ERROR nova.compute.manager [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d0ae7b46-9488-45b2-89ed-a66a603cdec1, please check neutron logs for more information. [ 696.150862] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 696.150862] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 696.150862] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 696.150862] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 696.150862] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 696.150862] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 696.150862] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 696.150862] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.150862] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 696.150862] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.150862] env[62600]: ERROR nova.compute.manager raise self.value [ 696.150862] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 696.150862] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 696.150862] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.150862] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 696.151519] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.151519] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 696.151519] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d0ae7b46-9488-45b2-89ed-a66a603cdec1, please check neutron logs for more information. [ 696.151519] env[62600]: ERROR nova.compute.manager [ 696.151519] env[62600]: Traceback (most recent call last): [ 696.151519] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 696.151519] env[62600]: listener.cb(fileno) [ 696.151519] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 696.151519] env[62600]: result = function(*args, **kwargs) [ 696.151519] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 696.151519] env[62600]: return func(*args, **kwargs) [ 696.151519] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 696.151519] env[62600]: raise e [ 696.151519] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 696.151519] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 696.151519] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 696.151519] env[62600]: created_port_ids = self._update_ports_for_instance( [ 696.151519] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 696.151519] env[62600]: with excutils.save_and_reraise_exception(): [ 696.151519] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.151519] env[62600]: self.force_reraise() [ 696.151519] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.151519] env[62600]: raise self.value [ 696.151519] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 696.151519] env[62600]: updated_port = self._update_port( [ 696.151519] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.151519] env[62600]: _ensure_no_port_binding_failure(port) [ 696.151519] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.151519] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 696.152270] env[62600]: nova.exception.PortBindingFailed: Binding failed for port d0ae7b46-9488-45b2-89ed-a66a603cdec1, please check neutron logs for more information. [ 696.152270] env[62600]: Removing descriptor: 15 [ 696.152270] env[62600]: ERROR nova.compute.manager [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d0ae7b46-9488-45b2-89ed-a66a603cdec1, please check neutron logs for more information. [ 696.152270] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Traceback (most recent call last): [ 696.152270] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 696.152270] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] yield resources [ 696.152270] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 696.152270] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] self.driver.spawn(context, instance, image_meta, [ 696.152270] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 696.152270] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 696.152270] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 696.152270] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] vm_ref = self.build_virtual_machine(instance, [ 696.152604] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 696.152604] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] vif_infos = vmwarevif.get_vif_info(self._session, [ 696.152604] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 696.152604] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] for vif in network_info: [ 696.152604] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 696.152604] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] return self._sync_wrapper(fn, *args, **kwargs) [ 696.152604] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 696.152604] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] self.wait() [ 696.152604] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 696.152604] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] self[:] = self._gt.wait() [ 696.152604] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 696.152604] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] return self._exit_event.wait() [ 696.152604] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 696.152956] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] result = hub.switch() [ 696.152956] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 696.152956] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] return self.greenlet.switch() [ 696.152956] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 696.152956] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] result = function(*args, **kwargs) [ 696.152956] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 696.152956] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] return func(*args, **kwargs) [ 696.152956] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 696.152956] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] raise e [ 696.152956] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 696.152956] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] nwinfo = self.network_api.allocate_for_instance( [ 696.152956] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 696.152956] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] created_port_ids = self._update_ports_for_instance( [ 696.153287] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 696.153287] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] with excutils.save_and_reraise_exception(): [ 696.153287] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.153287] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] self.force_reraise() [ 696.153287] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.153287] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] raise self.value [ 696.153287] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 696.153287] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] updated_port = self._update_port( [ 696.153287] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.153287] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] _ensure_no_port_binding_failure(port) [ 696.153287] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.153287] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] raise exception.PortBindingFailed(port_id=port['id']) [ 696.153565] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] nova.exception.PortBindingFailed: Binding failed for port d0ae7b46-9488-45b2-89ed-a66a603cdec1, please check neutron logs for more information. [ 696.153565] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] [ 696.153565] env[62600]: INFO nova.compute.manager [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Terminating instance [ 696.157849] env[62600]: DEBUG oslo_concurrency.lockutils [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Acquiring lock "refresh_cache-15772d8e-98ce-47cf-9b61-e79c7ffc3a8b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.455879] env[62600]: DEBUG nova.network.neutron [req-4da20fbb-b23a-4151-83da-2dc812f6d29c req-39ea7f43-781c-4e18-ba13-23fee528d746 service nova] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.504015] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 877f5b67-68f8-48b0-8e12-622d9b50a390 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 696.542609] env[62600]: DEBUG nova.network.neutron [req-4da20fbb-b23a-4151-83da-2dc812f6d29c req-39ea7f43-781c-4e18-ba13-23fee528d746 service nova] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.007309] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.048154] env[62600]: DEBUG oslo_concurrency.lockutils [req-4da20fbb-b23a-4151-83da-2dc812f6d29c req-39ea7f43-781c-4e18-ba13-23fee528d746 service nova] Releasing lock "refresh_cache-15772d8e-98ce-47cf-9b61-e79c7ffc3a8b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.048154] env[62600]: DEBUG oslo_concurrency.lockutils [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Acquired lock "refresh_cache-15772d8e-98ce-47cf-9b61-e79c7ffc3a8b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.048154] env[62600]: DEBUG nova.network.neutron [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 697.512191] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance f989a4e4-c3c6-4cb5-9464-cbfb9d66c202 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.569864] env[62600]: DEBUG nova.network.neutron [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.669622] env[62600]: DEBUG nova.network.neutron [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.017681] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 9c2c8dd4-85c7-489a-95c0-669644237b8b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 698.037327] env[62600]: DEBUG nova.compute.manager [req-e149f11c-39a8-4d7f-8a5e-e169a306655a req-3f7281e8-be70-4f2d-85da-fcac49d8e971 service nova] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Received event network-vif-deleted-d0ae7b46-9488-45b2-89ed-a66a603cdec1 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 698.172037] env[62600]: DEBUG oslo_concurrency.lockutils [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Releasing lock "refresh_cache-15772d8e-98ce-47cf-9b61-e79c7ffc3a8b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.172429] env[62600]: DEBUG nova.compute.manager [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 698.173949] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 698.173949] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ad281e6f-a009-4558-a0d8-14bd435d10bb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.187216] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32180c63-e13a-4ccd-9ad3-818439eec485 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.220575] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b could not be found. [ 698.220575] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 698.220575] env[62600]: INFO nova.compute.manager [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Took 0.05 seconds to destroy the instance on the hypervisor. [ 698.220575] env[62600]: DEBUG oslo.service.loopingcall [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 698.222034] env[62600]: DEBUG nova.compute.manager [-] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 698.222034] env[62600]: DEBUG nova.network.neutron [-] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 698.243392] env[62600]: DEBUG nova.network.neutron [-] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.523486] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 81814471-a617-4364-a140-105252ca9c04 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 698.747068] env[62600]: DEBUG nova.network.neutron [-] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.028837] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 699.249248] env[62600]: INFO nova.compute.manager [-] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Took 1.03 seconds to deallocate network for instance. [ 699.252688] env[62600]: DEBUG nova.compute.claims [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 699.252896] env[62600]: DEBUG oslo_concurrency.lockutils [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.532399] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 606becc8-ff43-469c-aca6-c35d67614960 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 700.039135] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 014899cb-db4c-4444-abe1-b0129c0d7db9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 700.335362] env[62600]: DEBUG oslo_concurrency.lockutils [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Acquiring lock "58bf5b37-05dd-478f-8c0b-5475ab63bf71" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.335362] env[62600]: DEBUG oslo_concurrency.lockutils [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Lock "58bf5b37-05dd-478f-8c0b-5475ab63bf71" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.543095] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance de4f9304-3357-4eaa-9c94-fe28bc554086 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 701.047552] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 92cf1acb-b471-4add-a73f-7e9e94fbaaa4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 701.549418] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance efff4d72-4673-4a8d-9a81-be9d700ff881 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 702.052873] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance a358a3cb-deda-419a-aa3c-ce7aeb534240 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 702.556093] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance e25c631d-3d1b-40d2-9fb1-a65431f991a9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 703.060460] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 703.564086] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance dcbe8e2d-6ae9-465f-8394-3978ee61a15b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 704.069034] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 704.572516] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance d34c098d-a827-4dbd-96e4-e27d9d56b847 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 705.075553] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 705.578640] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 734e8a2b-faaa-4165-b0e9-69eeb3699b2d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 705.579016] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 705.579196] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 705.931902] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b97bac4b-8520-4220-a523-6554f19c713e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.939898] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a474c020-a043-4041-9c46-d0ea00dd11d3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.969310] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feaa8a9c-0eaa-45ba-bed7-34d2ef075f79 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.976984] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3919b331-dccb-4cdf-9408-354440a15609 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.991550] env[62600]: DEBUG nova.compute.provider_tree [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.096184] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquiring lock "afe838cc-e086-4986-87ec-4e1266bcaf60" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.096453] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "afe838cc-e086-4986-87ec-4e1266bcaf60" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.494670] env[62600]: DEBUG nova.scheduler.client.report [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 706.999935] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62600) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 707.000035] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 13.068s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.000322] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.815s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.825961] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61062a90-f05f-4c80-8e84-70b63e33b42b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.833491] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a23445f-d578-469d-bcbe-edea8d278139 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.863631] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc3d5ff7-19d6-48da-9989-f4f2d6158442 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.869484] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-097ba2d0-3da5-4d5c-9c62-c4edead7e8ac {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.883373] env[62600]: DEBUG nova.compute.provider_tree [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.386950] env[62600]: DEBUG nova.scheduler.client.report [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 708.891772] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.891s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.892486] env[62600]: ERROR nova.compute.manager [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8f0a9399-ca7f-4eb3-b9f3-c5a1b6a20b46, please check neutron logs for more information. [ 708.892486] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Traceback (most recent call last): [ 708.892486] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 708.892486] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] self.driver.spawn(context, instance, image_meta, [ 708.892486] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 708.892486] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] self._vmops.spawn(context, instance, image_meta, injected_files, [ 708.892486] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 708.892486] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] vm_ref = self.build_virtual_machine(instance, [ 708.892486] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 708.892486] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] vif_infos = vmwarevif.get_vif_info(self._session, [ 708.892486] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 708.892816] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] for vif in network_info: [ 708.892816] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 708.892816] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] return self._sync_wrapper(fn, *args, **kwargs) [ 708.892816] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 708.892816] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] self.wait() [ 708.892816] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 708.892816] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] self[:] = self._gt.wait() [ 708.892816] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 708.892816] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] return self._exit_event.wait() [ 708.892816] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 708.892816] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] current.throw(*self._exc) [ 708.892816] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.892816] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] result = function(*args, **kwargs) [ 708.893136] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 708.893136] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] return func(*args, **kwargs) [ 708.893136] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 708.893136] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] raise e [ 708.893136] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.893136] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] nwinfo = self.network_api.allocate_for_instance( [ 708.893136] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 708.893136] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] created_port_ids = self._update_ports_for_instance( [ 708.893136] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 708.893136] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] with excutils.save_and_reraise_exception(): [ 708.893136] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.893136] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] self.force_reraise() [ 708.893136] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.893494] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] raise self.value [ 708.893494] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 708.893494] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] updated_port = self._update_port( [ 708.893494] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.893494] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] _ensure_no_port_binding_failure(port) [ 708.893494] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.893494] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] raise exception.PortBindingFailed(port_id=port['id']) [ 708.893494] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] nova.exception.PortBindingFailed: Binding failed for port 8f0a9399-ca7f-4eb3-b9f3-c5a1b6a20b46, please check neutron logs for more information. [ 708.893494] env[62600]: ERROR nova.compute.manager [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] [ 708.893494] env[62600]: DEBUG nova.compute.utils [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Binding failed for port 8f0a9399-ca7f-4eb3-b9f3-c5a1b6a20b46, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 708.894420] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.787s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.897328] env[62600]: DEBUG nova.compute.manager [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Build of instance e8cbb437-0ace-4af3-a1d9-f96eeeea8504 was re-scheduled: Binding failed for port 8f0a9399-ca7f-4eb3-b9f3-c5a1b6a20b46, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 708.897782] env[62600]: DEBUG nova.compute.manager [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 708.898011] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Acquiring lock "refresh_cache-e8cbb437-0ace-4af3-a1d9-f96eeeea8504" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.898143] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Acquired lock "refresh_cache-e8cbb437-0ace-4af3-a1d9-f96eeeea8504" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.898301] env[62600]: DEBUG nova.network.neutron [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 709.422898] env[62600]: DEBUG nova.network.neutron [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 709.509568] env[62600]: DEBUG nova.network.neutron [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.726528] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ea80e6-532d-42d2-8ee3-186483194542 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.735029] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9189fde-b262-4e38-8c43-54b284ab6417 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.764778] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9078fd73-7943-4d2d-bc13-5e8ad9ba4961 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.772133] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d3b62a-3681-4b64-b508-2e5407e90fc7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.786705] env[62600]: DEBUG nova.compute.provider_tree [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.014827] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Releasing lock "refresh_cache-e8cbb437-0ace-4af3-a1d9-f96eeeea8504" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.015188] env[62600]: DEBUG nova.compute.manager [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 710.015429] env[62600]: DEBUG nova.compute.manager [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 710.015630] env[62600]: DEBUG nova.network.neutron [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 710.031363] env[62600]: DEBUG nova.network.neutron [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 710.289867] env[62600]: DEBUG nova.scheduler.client.report [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.533768] env[62600]: DEBUG nova.network.neutron [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.797017] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.902s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.797136] env[62600]: ERROR nova.compute.manager [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 15dceed5-8643-4f6b-ba3c-0ec3b34f0684, please check neutron logs for more information. [ 710.797136] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] Traceback (most recent call last): [ 710.797136] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 710.797136] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] self.driver.spawn(context, instance, image_meta, [ 710.797136] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 710.797136] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] self._vmops.spawn(context, instance, image_meta, injected_files, [ 710.797136] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 710.797136] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] vm_ref = self.build_virtual_machine(instance, [ 710.797136] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 710.797136] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] vif_infos = vmwarevif.get_vif_info(self._session, [ 710.797136] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 710.797475] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] for vif in network_info: [ 710.797475] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 710.797475] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] return self._sync_wrapper(fn, *args, **kwargs) [ 710.797475] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 710.797475] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] self.wait() [ 710.797475] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 710.797475] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] self[:] = self._gt.wait() [ 710.797475] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 710.797475] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] return self._exit_event.wait() [ 710.797475] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 710.797475] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] current.throw(*self._exc) [ 710.797475] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.797475] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] result = function(*args, **kwargs) [ 710.797844] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 710.797844] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] return func(*args, **kwargs) [ 710.797844] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.797844] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] raise e [ 710.797844] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.797844] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] nwinfo = self.network_api.allocate_for_instance( [ 710.797844] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 710.797844] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] created_port_ids = self._update_ports_for_instance( [ 710.797844] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 710.797844] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] with excutils.save_and_reraise_exception(): [ 710.797844] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.797844] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] self.force_reraise() [ 710.797844] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.798225] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] raise self.value [ 710.798225] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 710.798225] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] updated_port = self._update_port( [ 710.798225] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.798225] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] _ensure_no_port_binding_failure(port) [ 710.798225] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.798225] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] raise exception.PortBindingFailed(port_id=port['id']) [ 710.798225] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] nova.exception.PortBindingFailed: Binding failed for port 15dceed5-8643-4f6b-ba3c-0ec3b34f0684, please check neutron logs for more information. [ 710.798225] env[62600]: ERROR nova.compute.manager [instance: 3d19d349-e450-4307-a763-6269dcb06544] [ 710.798225] env[62600]: DEBUG nova.compute.utils [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Binding failed for port 15dceed5-8643-4f6b-ba3c-0ec3b34f0684, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 710.798969] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.211s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.801767] env[62600]: DEBUG nova.compute.manager [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Build of instance 3d19d349-e450-4307-a763-6269dcb06544 was re-scheduled: Binding failed for port 15dceed5-8643-4f6b-ba3c-0ec3b34f0684, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 710.802185] env[62600]: DEBUG nova.compute.manager [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 710.802406] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Acquiring lock "refresh_cache-3d19d349-e450-4307-a763-6269dcb06544" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.802550] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Acquired lock "refresh_cache-3d19d349-e450-4307-a763-6269dcb06544" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.802706] env[62600]: DEBUG nova.network.neutron [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 711.037049] env[62600]: INFO nova.compute.manager [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] [instance: e8cbb437-0ace-4af3-a1d9-f96eeeea8504] Took 1.02 seconds to deallocate network for instance. [ 711.329217] env[62600]: DEBUG nova.network.neutron [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 711.405920] env[62600]: DEBUG nova.network.neutron [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.628500] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad767d2-c86b-4b78-a3ee-b93b8f96fce5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.635850] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06162d51-c406-4dbe-a29e-b67472f96d65 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.666051] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0e04fa3-7a7c-4cde-b4e9-ce9130bcc6b0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.672690] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69a8a2dd-019c-462c-8f6d-bb006c98dfd3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.685247] env[62600]: DEBUG nova.compute.provider_tree [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.908650] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Releasing lock "refresh_cache-3d19d349-e450-4307-a763-6269dcb06544" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.908901] env[62600]: DEBUG nova.compute.manager [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 711.909186] env[62600]: DEBUG nova.compute.manager [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 711.909359] env[62600]: DEBUG nova.network.neutron [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 711.931964] env[62600]: DEBUG nova.network.neutron [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 712.070252] env[62600]: INFO nova.scheduler.client.report [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Deleted allocations for instance e8cbb437-0ace-4af3-a1d9-f96eeeea8504 [ 712.188574] env[62600]: DEBUG nova.scheduler.client.report [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.436294] env[62600]: DEBUG nova.network.neutron [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.577388] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b414a008-4bd3-4b58-8bb2-f770e11ad91f tempest-ServerActionsTestOtherA-1837948896 tempest-ServerActionsTestOtherA-1837948896-project-member] Lock "e8cbb437-0ace-4af3-a1d9-f96eeeea8504" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.650s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.694195] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.895s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.694927] env[62600]: ERROR nova.compute.manager [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cfd31541-18d4-4711-80f4-ee76d04de9e9, please check neutron logs for more information. [ 712.694927] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Traceback (most recent call last): [ 712.694927] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.694927] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] self.driver.spawn(context, instance, image_meta, [ 712.694927] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 712.694927] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.694927] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.694927] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] vm_ref = self.build_virtual_machine(instance, [ 712.694927] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.694927] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.694927] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.695303] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] for vif in network_info: [ 712.695303] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.695303] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] return self._sync_wrapper(fn, *args, **kwargs) [ 712.695303] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.695303] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] self.wait() [ 712.695303] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.695303] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] self[:] = self._gt.wait() [ 712.695303] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.695303] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] return self._exit_event.wait() [ 712.695303] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 712.695303] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] current.throw(*self._exc) [ 712.695303] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.695303] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] result = function(*args, **kwargs) [ 712.695740] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 712.695740] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] return func(*args, **kwargs) [ 712.695740] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.695740] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] raise e [ 712.695740] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.695740] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] nwinfo = self.network_api.allocate_for_instance( [ 712.695740] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 712.695740] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] created_port_ids = self._update_ports_for_instance( [ 712.695740] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 712.695740] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] with excutils.save_and_reraise_exception(): [ 712.695740] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.695740] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] self.force_reraise() [ 712.695740] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.696125] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] raise self.value [ 712.696125] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 712.696125] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] updated_port = self._update_port( [ 712.696125] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.696125] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] _ensure_no_port_binding_failure(port) [ 712.696125] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.696125] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] raise exception.PortBindingFailed(port_id=port['id']) [ 712.696125] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] nova.exception.PortBindingFailed: Binding failed for port cfd31541-18d4-4711-80f4-ee76d04de9e9, please check neutron logs for more information. [ 712.696125] env[62600]: ERROR nova.compute.manager [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] [ 712.696125] env[62600]: DEBUG nova.compute.utils [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Binding failed for port cfd31541-18d4-4711-80f4-ee76d04de9e9, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 712.697585] env[62600]: DEBUG nova.compute.manager [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Build of instance 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6 was re-scheduled: Binding failed for port cfd31541-18d4-4711-80f4-ee76d04de9e9, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 712.697951] env[62600]: DEBUG nova.compute.manager [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 712.698202] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Acquiring lock "refresh_cache-932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.698350] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Acquired lock "refresh_cache-932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.698510] env[62600]: DEBUG nova.network.neutron [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 712.700024] env[62600]: DEBUG oslo_concurrency.lockutils [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.083s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.939191] env[62600]: INFO nova.compute.manager [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] [instance: 3d19d349-e450-4307-a763-6269dcb06544] Took 1.03 seconds to deallocate network for instance. [ 713.080669] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 713.226050] env[62600]: DEBUG nova.network.neutron [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 713.307088] env[62600]: DEBUG nova.network.neutron [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.605809] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.640430] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9553d271-b432-427f-83a5-58cce3b82564 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.648751] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0e904d-fb9e-4d3b-a985-0ffb61cc2a61 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.679336] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6240c099-cf16-4a66-bcc1-5db0f56f6f48 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.686697] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28cac968-c599-4953-aa6d-936cacc3a6aa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.699631] env[62600]: DEBUG nova.compute.provider_tree [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.813730] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Releasing lock "refresh_cache-932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.813978] env[62600]: DEBUG nova.compute.manager [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 713.814177] env[62600]: DEBUG nova.compute.manager [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 713.814349] env[62600]: DEBUG nova.network.neutron [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 713.832551] env[62600]: DEBUG nova.network.neutron [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 713.984735] env[62600]: INFO nova.scheduler.client.report [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Deleted allocations for instance 3d19d349-e450-4307-a763-6269dcb06544 [ 714.204472] env[62600]: DEBUG nova.scheduler.client.report [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.341953] env[62600]: DEBUG nova.network.neutron [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.498032] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12aaad3b-58c6-48e8-b25c-008c806a99d9 tempest-ImagesOneServerNegativeTestJSON-501481936 tempest-ImagesOneServerNegativeTestJSON-501481936-project-member] Lock "3d19d349-e450-4307-a763-6269dcb06544" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.777s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.708021] env[62600]: DEBUG oslo_concurrency.lockutils [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.008s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.708807] env[62600]: ERROR nova.compute.manager [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 776c7599-bde8-4d84-9c07-5796cc622714, please check neutron logs for more information. [ 714.708807] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Traceback (most recent call last): [ 714.708807] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 714.708807] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] self.driver.spawn(context, instance, image_meta, [ 714.708807] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 714.708807] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 714.708807] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 714.708807] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] vm_ref = self.build_virtual_machine(instance, [ 714.708807] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 714.708807] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 714.708807] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 714.709101] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] for vif in network_info: [ 714.709101] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 714.709101] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] return self._sync_wrapper(fn, *args, **kwargs) [ 714.709101] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 714.709101] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] self.wait() [ 714.709101] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 714.709101] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] self[:] = self._gt.wait() [ 714.709101] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 714.709101] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] return self._exit_event.wait() [ 714.709101] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 714.709101] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] result = hub.switch() [ 714.709101] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 714.709101] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] return self.greenlet.switch() [ 714.709450] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.709450] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] result = function(*args, **kwargs) [ 714.709450] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 714.709450] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] return func(*args, **kwargs) [ 714.709450] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.709450] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] raise e [ 714.709450] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.709450] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] nwinfo = self.network_api.allocate_for_instance( [ 714.709450] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 714.709450] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] created_port_ids = self._update_ports_for_instance( [ 714.709450] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 714.709450] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] with excutils.save_and_reraise_exception(): [ 714.709450] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.709877] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] self.force_reraise() [ 714.709877] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.709877] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] raise self.value [ 714.709877] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 714.709877] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] updated_port = self._update_port( [ 714.709877] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.709877] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] _ensure_no_port_binding_failure(port) [ 714.709877] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.709877] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] raise exception.PortBindingFailed(port_id=port['id']) [ 714.709877] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] nova.exception.PortBindingFailed: Binding failed for port 776c7599-bde8-4d84-9c07-5796cc622714, please check neutron logs for more information. [ 714.709877] env[62600]: ERROR nova.compute.manager [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] [ 714.710168] env[62600]: DEBUG nova.compute.utils [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Binding failed for port 776c7599-bde8-4d84-9c07-5796cc622714, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 714.711436] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.075s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.713395] env[62600]: INFO nova.compute.claims [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 714.716403] env[62600]: DEBUG nova.compute.manager [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Build of instance 20fab902-79cc-4f83-abaf-53a53c2521f4 was re-scheduled: Binding failed for port 776c7599-bde8-4d84-9c07-5796cc622714, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 714.717028] env[62600]: DEBUG nova.compute.manager [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 714.717242] env[62600]: DEBUG oslo_concurrency.lockutils [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Acquiring lock "refresh_cache-20fab902-79cc-4f83-abaf-53a53c2521f4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.717396] env[62600]: DEBUG oslo_concurrency.lockutils [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Acquired lock "refresh_cache-20fab902-79cc-4f83-abaf-53a53c2521f4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.717716] env[62600]: DEBUG nova.network.neutron [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 714.843347] env[62600]: INFO nova.compute.manager [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] [instance: 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6] Took 1.03 seconds to deallocate network for instance. [ 715.000718] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 715.245422] env[62600]: DEBUG nova.network.neutron [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 715.337080] env[62600]: DEBUG nova.network.neutron [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.526522] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.844913] env[62600]: DEBUG oslo_concurrency.lockutils [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Releasing lock "refresh_cache-20fab902-79cc-4f83-abaf-53a53c2521f4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.845162] env[62600]: DEBUG nova.compute.manager [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 715.845347] env[62600]: DEBUG nova.compute.manager [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 715.845513] env[62600]: DEBUG nova.network.neutron [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 715.868236] env[62600]: DEBUG nova.network.neutron [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 715.894278] env[62600]: INFO nova.scheduler.client.report [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Deleted allocations for instance 932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6 [ 716.123658] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba23b89-ab43-45eb-8a02-f95c4a30e7b0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.131411] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b816ce86-f670-4cf7-9304-4a383f659dee {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.162402] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c59aa0ed-379a-4913-a1da-11639b4bb10a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.169608] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b16fd020-54b6-4e73-b630-518d79f2217c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.186560] env[62600]: DEBUG nova.compute.provider_tree [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.370989] env[62600]: DEBUG nova.network.neutron [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.406460] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7ea250e0-bdbc-4182-b3be-4dbedd35a2d3 tempest-ServerAddressesNegativeTestJSON-650328106 tempest-ServerAddressesNegativeTestJSON-650328106-project-member] Lock "932b7ac4-0a4b-4a39-a002-5a4ae3ce75f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.671s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.691906] env[62600]: DEBUG nova.scheduler.client.report [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.874727] env[62600]: INFO nova.compute.manager [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] [instance: 20fab902-79cc-4f83-abaf-53a53c2521f4] Took 1.03 seconds to deallocate network for instance. [ 716.911453] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 717.198878] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.487s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.199439] env[62600]: DEBUG nova.compute.manager [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 717.203433] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.651s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.204253] env[62600]: INFO nova.compute.claims [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 717.445829] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.714539] env[62600]: DEBUG nova.compute.utils [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 717.716072] env[62600]: DEBUG nova.compute.manager [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 717.716255] env[62600]: DEBUG nova.network.neutron [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 717.824481] env[62600]: DEBUG nova.policy [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34319d68370c48ef93fa24a9d9132407', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bab74e3ae78248909dbcd483abbb8da2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 717.919923] env[62600]: INFO nova.scheduler.client.report [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Deleted allocations for instance 20fab902-79cc-4f83-abaf-53a53c2521f4 [ 718.223990] env[62600]: DEBUG nova.compute.manager [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 718.424159] env[62600]: DEBUG nova.network.neutron [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Successfully created port: ad089818-0871-43ff-8fd3-58654980f370 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 718.434779] env[62600]: DEBUG oslo_concurrency.lockutils [None req-43c9eec5-0ae7-4eb9-872b-7ceecdabbd48 tempest-ServersTestManualDisk-2003579840 tempest-ServersTestManualDisk-2003579840-project-member] Lock "20fab902-79cc-4f83-abaf-53a53c2521f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.449s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.593334] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c8969c-e920-4580-8309-d8d14d49402f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.605468] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e684733d-2be1-4ed1-90c5-11ec9805d968 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.643406] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a97872ae-dcfd-4542-a67c-05766ac3b12f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.651751] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebc0c261-5036-4a35-b001-4592dab2a07f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.666339] env[62600]: DEBUG nova.compute.provider_tree [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.938556] env[62600]: DEBUG nova.compute.manager [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 719.171104] env[62600]: DEBUG nova.scheduler.client.report [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 719.237016] env[62600]: DEBUG nova.compute.manager [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 719.263115] env[62600]: DEBUG nova.virt.hardware [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 719.263115] env[62600]: DEBUG nova.virt.hardware [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 719.263780] env[62600]: DEBUG nova.virt.hardware [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 719.264187] env[62600]: DEBUG nova.virt.hardware [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 719.264583] env[62600]: DEBUG nova.virt.hardware [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 719.264846] env[62600]: DEBUG nova.virt.hardware [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 719.265184] env[62600]: DEBUG nova.virt.hardware [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 719.265453] env[62600]: DEBUG nova.virt.hardware [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 719.265802] env[62600]: DEBUG nova.virt.hardware [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 719.266108] env[62600]: DEBUG nova.virt.hardware [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 719.266439] env[62600]: DEBUG nova.virt.hardware [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 719.267855] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f10576d6-7d0b-4bfb-a631-f56284508293 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.279232] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc2f754f-9d07-431a-a4e2-98f0caae5733 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.434482] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Acquiring lock "8f3e87a5-bb70-4e50-9fea-6bf53946b8dd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.434784] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Lock "8f3e87a5-bb70-4e50-9fea-6bf53946b8dd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.468644] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.676815] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.475s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.677387] env[62600]: DEBUG nova.compute.manager [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 719.684017] env[62600]: DEBUG oslo_concurrency.lockutils [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.958s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.919544] env[62600]: DEBUG nova.compute.manager [req-096de55a-67ab-44a3-a952-7a6265e4f017 req-37b46ae2-7ce2-493e-83ec-4e596fa5a485 service nova] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Received event network-changed-ad089818-0871-43ff-8fd3-58654980f370 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 719.923288] env[62600]: DEBUG nova.compute.manager [req-096de55a-67ab-44a3-a952-7a6265e4f017 req-37b46ae2-7ce2-493e-83ec-4e596fa5a485 service nova] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Refreshing instance network info cache due to event network-changed-ad089818-0871-43ff-8fd3-58654980f370. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 719.923288] env[62600]: DEBUG oslo_concurrency.lockutils [req-096de55a-67ab-44a3-a952-7a6265e4f017 req-37b46ae2-7ce2-493e-83ec-4e596fa5a485 service nova] Acquiring lock "refresh_cache-c93cb228-0adb-423c-854f-73354cdb7012" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.923288] env[62600]: DEBUG oslo_concurrency.lockutils [req-096de55a-67ab-44a3-a952-7a6265e4f017 req-37b46ae2-7ce2-493e-83ec-4e596fa5a485 service nova] Acquired lock "refresh_cache-c93cb228-0adb-423c-854f-73354cdb7012" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.923288] env[62600]: DEBUG nova.network.neutron [req-096de55a-67ab-44a3-a952-7a6265e4f017 req-37b46ae2-7ce2-493e-83ec-4e596fa5a485 service nova] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Refreshing network info cache for port ad089818-0871-43ff-8fd3-58654980f370 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 720.190031] env[62600]: DEBUG nova.compute.utils [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 720.195036] env[62600]: DEBUG nova.compute.manager [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 720.195564] env[62600]: DEBUG nova.network.neutron [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 720.237811] env[62600]: ERROR nova.compute.manager [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ad089818-0871-43ff-8fd3-58654980f370, please check neutron logs for more information. [ 720.237811] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 720.237811] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.237811] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 720.237811] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 720.237811] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 720.237811] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 720.237811] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 720.237811] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.237811] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 720.237811] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.237811] env[62600]: ERROR nova.compute.manager raise self.value [ 720.237811] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 720.237811] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 720.237811] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.237811] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 720.238399] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.238399] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 720.238399] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ad089818-0871-43ff-8fd3-58654980f370, please check neutron logs for more information. [ 720.238399] env[62600]: ERROR nova.compute.manager [ 720.238399] env[62600]: Traceback (most recent call last): [ 720.238399] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 720.238399] env[62600]: listener.cb(fileno) [ 720.238399] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.238399] env[62600]: result = function(*args, **kwargs) [ 720.238399] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 720.238399] env[62600]: return func(*args, **kwargs) [ 720.238399] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.238399] env[62600]: raise e [ 720.238399] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.238399] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 720.238399] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 720.238399] env[62600]: created_port_ids = self._update_ports_for_instance( [ 720.238399] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 720.238399] env[62600]: with excutils.save_and_reraise_exception(): [ 720.238399] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.238399] env[62600]: self.force_reraise() [ 720.238399] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.238399] env[62600]: raise self.value [ 720.238399] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 720.238399] env[62600]: updated_port = self._update_port( [ 720.238399] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.238399] env[62600]: _ensure_no_port_binding_failure(port) [ 720.238399] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.238399] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 720.239175] env[62600]: nova.exception.PortBindingFailed: Binding failed for port ad089818-0871-43ff-8fd3-58654980f370, please check neutron logs for more information. [ 720.239175] env[62600]: Removing descriptor: 15 [ 720.239175] env[62600]: ERROR nova.compute.manager [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ad089818-0871-43ff-8fd3-58654980f370, please check neutron logs for more information. [ 720.239175] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] Traceback (most recent call last): [ 720.239175] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 720.239175] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] yield resources [ 720.239175] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 720.239175] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] self.driver.spawn(context, instance, image_meta, [ 720.239175] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 720.239175] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] self._vmops.spawn(context, instance, image_meta, injected_files, [ 720.239175] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 720.239175] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] vm_ref = self.build_virtual_machine(instance, [ 720.239536] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 720.239536] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] vif_infos = vmwarevif.get_vif_info(self._session, [ 720.239536] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 720.239536] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] for vif in network_info: [ 720.239536] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 720.239536] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] return self._sync_wrapper(fn, *args, **kwargs) [ 720.239536] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 720.239536] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] self.wait() [ 720.239536] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 720.239536] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] self[:] = self._gt.wait() [ 720.239536] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 720.239536] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] return self._exit_event.wait() [ 720.239536] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 720.239918] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] result = hub.switch() [ 720.239918] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 720.239918] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] return self.greenlet.switch() [ 720.239918] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.239918] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] result = function(*args, **kwargs) [ 720.239918] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 720.239918] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] return func(*args, **kwargs) [ 720.239918] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 720.239918] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] raise e [ 720.239918] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 720.239918] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] nwinfo = self.network_api.allocate_for_instance( [ 720.239918] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 720.239918] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] created_port_ids = self._update_ports_for_instance( [ 720.240485] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 720.240485] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] with excutils.save_and_reraise_exception(): [ 720.240485] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.240485] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] self.force_reraise() [ 720.240485] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.240485] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] raise self.value [ 720.240485] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 720.240485] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] updated_port = self._update_port( [ 720.240485] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.240485] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] _ensure_no_port_binding_failure(port) [ 720.240485] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.240485] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] raise exception.PortBindingFailed(port_id=port['id']) [ 720.240840] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] nova.exception.PortBindingFailed: Binding failed for port ad089818-0871-43ff-8fd3-58654980f370, please check neutron logs for more information. [ 720.240840] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] [ 720.240840] env[62600]: INFO nova.compute.manager [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Terminating instance [ 720.244659] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "refresh_cache-c93cb228-0adb-423c-854f-73354cdb7012" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.273524] env[62600]: DEBUG nova.policy [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e815d597c84a4048920ba11bbac927ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '03da7a5233f4486b8e5b08ea12776f97', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 720.450442] env[62600]: DEBUG nova.network.neutron [req-096de55a-67ab-44a3-a952-7a6265e4f017 req-37b46ae2-7ce2-493e-83ec-4e596fa5a485 service nova] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 720.639220] env[62600]: DEBUG nova.network.neutron [req-096de55a-67ab-44a3-a952-7a6265e4f017 req-37b46ae2-7ce2-493e-83ec-4e596fa5a485 service nova] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.647319] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba16f335-fcdb-4ff4-b77c-fa2179219c29 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.657155] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e09ddb58-2d69-44b4-a9fa-ccb826f012f5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.692256] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9abaec12-4cb3-4397-a38e-c72d6200baef {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.695144] env[62600]: DEBUG nova.compute.manager [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 720.706020] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab4493c1-51d1-47a2-b6d8-cd65c80dced2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.722198] env[62600]: DEBUG nova.compute.provider_tree [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.844987] env[62600]: DEBUG nova.network.neutron [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Successfully created port: 7714dcfd-c51e-4fd3-ae24-8fed6b64d011 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 721.145741] env[62600]: DEBUG oslo_concurrency.lockutils [req-096de55a-67ab-44a3-a952-7a6265e4f017 req-37b46ae2-7ce2-493e-83ec-4e596fa5a485 service nova] Releasing lock "refresh_cache-c93cb228-0adb-423c-854f-73354cdb7012" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.146247] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired lock "refresh_cache-c93cb228-0adb-423c-854f-73354cdb7012" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.146433] env[62600]: DEBUG nova.network.neutron [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 721.226123] env[62600]: DEBUG nova.scheduler.client.report [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 721.675031] env[62600]: DEBUG nova.network.neutron [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 721.712231] env[62600]: DEBUG nova.compute.manager [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 721.732521] env[62600]: DEBUG oslo_concurrency.lockutils [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.051s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.733849] env[62600]: ERROR nova.compute.manager [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e97d0860-1801-4d1e-b455-62b34fa87099, please check neutron logs for more information. [ 721.733849] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] Traceback (most recent call last): [ 721.733849] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 721.733849] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] self.driver.spawn(context, instance, image_meta, [ 721.733849] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 721.733849] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] self._vmops.spawn(context, instance, image_meta, injected_files, [ 721.733849] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 721.733849] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] vm_ref = self.build_virtual_machine(instance, [ 721.733849] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 721.733849] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] vif_infos = vmwarevif.get_vif_info(self._session, [ 721.733849] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 721.734257] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] for vif in network_info: [ 721.734257] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 721.734257] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] return self._sync_wrapper(fn, *args, **kwargs) [ 721.734257] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 721.734257] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] self.wait() [ 721.734257] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 721.734257] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] self[:] = self._gt.wait() [ 721.734257] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 721.734257] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] return self._exit_event.wait() [ 721.734257] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 721.734257] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] current.throw(*self._exc) [ 721.734257] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.734257] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] result = function(*args, **kwargs) [ 721.734633] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 721.734633] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] return func(*args, **kwargs) [ 721.734633] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.734633] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] raise e [ 721.734633] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.734633] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] nwinfo = self.network_api.allocate_for_instance( [ 721.734633] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 721.734633] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] created_port_ids = self._update_ports_for_instance( [ 721.734633] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 721.734633] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] with excutils.save_and_reraise_exception(): [ 721.734633] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.734633] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] self.force_reraise() [ 721.734633] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.734970] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] raise self.value [ 721.734970] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 721.734970] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] updated_port = self._update_port( [ 721.734970] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.734970] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] _ensure_no_port_binding_failure(port) [ 721.734970] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.734970] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] raise exception.PortBindingFailed(port_id=port['id']) [ 721.734970] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] nova.exception.PortBindingFailed: Binding failed for port e97d0860-1801-4d1e-b455-62b34fa87099, please check neutron logs for more information. [ 721.734970] env[62600]: ERROR nova.compute.manager [instance: 6b19c327-9a80-498f-a350-8068ad250377] [ 721.734970] env[62600]: DEBUG nova.compute.utils [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Binding failed for port e97d0860-1801-4d1e-b455-62b34fa87099, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 721.737569] env[62600]: DEBUG oslo_concurrency.lockutils [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.636s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.739527] env[62600]: INFO nova.compute.claims [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 721.745699] env[62600]: DEBUG nova.compute.manager [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Build of instance 6b19c327-9a80-498f-a350-8068ad250377 was re-scheduled: Binding failed for port e97d0860-1801-4d1e-b455-62b34fa87099, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 721.746583] env[62600]: DEBUG nova.compute.manager [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 721.746583] env[62600]: DEBUG oslo_concurrency.lockutils [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquiring lock "refresh_cache-6b19c327-9a80-498f-a350-8068ad250377" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.746583] env[62600]: DEBUG oslo_concurrency.lockutils [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquired lock "refresh_cache-6b19c327-9a80-498f-a350-8068ad250377" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.746968] env[62600]: DEBUG nova.network.neutron [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 721.760231] env[62600]: DEBUG nova.virt.hardware [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 721.760231] env[62600]: DEBUG nova.virt.hardware [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 721.760231] env[62600]: DEBUG nova.virt.hardware [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 721.760454] env[62600]: DEBUG nova.virt.hardware [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 721.760454] env[62600]: DEBUG nova.virt.hardware [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 721.761542] env[62600]: DEBUG nova.virt.hardware [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 721.761542] env[62600]: DEBUG nova.virt.hardware [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 721.761542] env[62600]: DEBUG nova.virt.hardware [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 721.761542] env[62600]: DEBUG nova.virt.hardware [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 721.761542] env[62600]: DEBUG nova.virt.hardware [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 721.761712] env[62600]: DEBUG nova.virt.hardware [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 721.762768] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-049e276f-856e-47c1-b9a5-9ed3fa974098 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.777098] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3f67e65-4141-49cf-88f9-fc50dcab0dc9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.852837] env[62600]: DEBUG nova.network.neutron [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.970718] env[62600]: DEBUG nova.compute.manager [req-30c26297-c8e4-4514-ae3b-650c2e3623b1 req-8740ab42-ee9f-43ee-bb70-39b8358c9342 service nova] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Received event network-vif-deleted-ad089818-0871-43ff-8fd3-58654980f370 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 722.274927] env[62600]: DEBUG nova.network.neutron [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 722.340565] env[62600]: ERROR nova.compute.manager [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7714dcfd-c51e-4fd3-ae24-8fed6b64d011, please check neutron logs for more information. [ 722.340565] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 722.340565] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.340565] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 722.340565] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 722.340565] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 722.340565] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 722.340565] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 722.340565] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.340565] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 722.340565] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.340565] env[62600]: ERROR nova.compute.manager raise self.value [ 722.340565] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 722.340565] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 722.340565] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.340565] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 722.341073] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.341073] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 722.341073] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7714dcfd-c51e-4fd3-ae24-8fed6b64d011, please check neutron logs for more information. [ 722.341073] env[62600]: ERROR nova.compute.manager [ 722.341880] env[62600]: Traceback (most recent call last): [ 722.341937] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 722.341937] env[62600]: listener.cb(fileno) [ 722.341937] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.341937] env[62600]: result = function(*args, **kwargs) [ 722.341937] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 722.341937] env[62600]: return func(*args, **kwargs) [ 722.341937] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 722.341937] env[62600]: raise e [ 722.341937] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.341937] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 722.341937] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 722.341937] env[62600]: created_port_ids = self._update_ports_for_instance( [ 722.341937] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 722.341937] env[62600]: with excutils.save_and_reraise_exception(): [ 722.341937] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.341937] env[62600]: self.force_reraise() [ 722.341937] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.341937] env[62600]: raise self.value [ 722.341937] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 722.341937] env[62600]: updated_port = self._update_port( [ 722.341937] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.341937] env[62600]: _ensure_no_port_binding_failure(port) [ 722.341937] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.341937] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 722.341937] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 7714dcfd-c51e-4fd3-ae24-8fed6b64d011, please check neutron logs for more information. [ 722.342572] env[62600]: Removing descriptor: 16 [ 722.342717] env[62600]: ERROR nova.compute.manager [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7714dcfd-c51e-4fd3-ae24-8fed6b64d011, please check neutron logs for more information. [ 722.342717] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Traceback (most recent call last): [ 722.342717] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 722.342717] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] yield resources [ 722.342717] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 722.342717] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] self.driver.spawn(context, instance, image_meta, [ 722.342717] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 722.342717] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] self._vmops.spawn(context, instance, image_meta, injected_files, [ 722.342717] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 722.342717] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] vm_ref = self.build_virtual_machine(instance, [ 722.342717] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 722.343036] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] vif_infos = vmwarevif.get_vif_info(self._session, [ 722.343036] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 722.343036] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] for vif in network_info: [ 722.343036] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 722.343036] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] return self._sync_wrapper(fn, *args, **kwargs) [ 722.343036] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 722.343036] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] self.wait() [ 722.343036] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 722.343036] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] self[:] = self._gt.wait() [ 722.343036] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 722.343036] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] return self._exit_event.wait() [ 722.343036] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 722.343036] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] result = hub.switch() [ 722.343362] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 722.343362] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] return self.greenlet.switch() [ 722.343362] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.343362] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] result = function(*args, **kwargs) [ 722.343362] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 722.343362] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] return func(*args, **kwargs) [ 722.343362] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 722.343362] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] raise e [ 722.343362] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.343362] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] nwinfo = self.network_api.allocate_for_instance( [ 722.343362] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 722.343362] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] created_port_ids = self._update_ports_for_instance( [ 722.343362] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 722.343697] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] with excutils.save_and_reraise_exception(): [ 722.343697] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.343697] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] self.force_reraise() [ 722.343697] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.343697] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] raise self.value [ 722.343697] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 722.343697] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] updated_port = self._update_port( [ 722.343697] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.343697] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] _ensure_no_port_binding_failure(port) [ 722.343697] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.343697] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] raise exception.PortBindingFailed(port_id=port['id']) [ 722.343697] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] nova.exception.PortBindingFailed: Binding failed for port 7714dcfd-c51e-4fd3-ae24-8fed6b64d011, please check neutron logs for more information. [ 722.343697] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] [ 722.344050] env[62600]: INFO nova.compute.manager [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Terminating instance [ 722.352160] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Acquiring lock "refresh_cache-e3416076-1d11-47e4-9f76-9a64d7259166" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.352413] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Acquired lock "refresh_cache-e3416076-1d11-47e4-9f76-9a64d7259166" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.353023] env[62600]: DEBUG nova.network.neutron [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 722.354992] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Releasing lock "refresh_cache-c93cb228-0adb-423c-854f-73354cdb7012" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.355563] env[62600]: DEBUG nova.compute.manager [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 722.355835] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 722.356185] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-04998e9c-4549-43bd-9260-ac933ef2e789 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.365603] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e91131f2-cd81-4f65-ad1b-f45405142f59 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.392335] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c93cb228-0adb-423c-854f-73354cdb7012 could not be found. [ 722.392716] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 722.393127] env[62600]: INFO nova.compute.manager [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Took 0.04 seconds to destroy the instance on the hypervisor. [ 722.393453] env[62600]: DEBUG oslo.service.loopingcall [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 722.394264] env[62600]: DEBUG nova.network.neutron [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.395733] env[62600]: DEBUG nova.compute.manager [-] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.395864] env[62600]: DEBUG nova.network.neutron [-] [instance: c93cb228-0adb-423c-854f-73354cdb7012] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 722.411043] env[62600]: DEBUG nova.network.neutron [-] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 722.879945] env[62600]: DEBUG nova.network.neutron [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 722.898991] env[62600]: DEBUG oslo_concurrency.lockutils [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Releasing lock "refresh_cache-6b19c327-9a80-498f-a350-8068ad250377" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.899550] env[62600]: DEBUG nova.compute.manager [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 722.899883] env[62600]: DEBUG nova.compute.manager [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.899929] env[62600]: DEBUG nova.network.neutron [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 722.913984] env[62600]: DEBUG nova.network.neutron [-] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.922105] env[62600]: DEBUG nova.network.neutron [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 722.989334] env[62600]: DEBUG nova.network.neutron [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.146814] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f1c88fb-bbd7-47e8-a158-4d51ba02f2e8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.155585] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d822b6c8-9fca-48e4-89d3-a8bfbfadf628 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.184706] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91356f5f-dd5b-4edf-8705-c36c7f5c337c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.192132] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd2fd1e9-6374-4c62-b970-b48cf111ed2b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.206962] env[62600]: DEBUG nova.compute.provider_tree [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.378870] env[62600]: DEBUG oslo_concurrency.lockutils [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "bd9b7351-01a2-429d-a860-d9ff6855eefc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.379367] env[62600]: DEBUG oslo_concurrency.lockutils [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "bd9b7351-01a2-429d-a860-d9ff6855eefc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.415999] env[62600]: INFO nova.compute.manager [-] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Took 1.02 seconds to deallocate network for instance. [ 723.418213] env[62600]: DEBUG nova.compute.claims [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 723.418388] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.424835] env[62600]: DEBUG nova.network.neutron [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.492074] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Releasing lock "refresh_cache-e3416076-1d11-47e4-9f76-9a64d7259166" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.492534] env[62600]: DEBUG nova.compute.manager [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 723.492729] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 723.493064] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fc8c6078-51c2-4fee-8d1c-327fc9a9d038 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.501861] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c217abb-b909-457b-8488-1a780430f1bf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.524242] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e3416076-1d11-47e4-9f76-9a64d7259166 could not be found. [ 723.524465] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 723.524646] env[62600]: INFO nova.compute.manager [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Took 0.03 seconds to destroy the instance on the hypervisor. [ 723.524887] env[62600]: DEBUG oslo.service.loopingcall [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 723.525118] env[62600]: DEBUG nova.compute.manager [-] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 723.525219] env[62600]: DEBUG nova.network.neutron [-] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 723.541021] env[62600]: DEBUG nova.network.neutron [-] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 723.712279] env[62600]: DEBUG nova.scheduler.client.report [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 723.930177] env[62600]: INFO nova.compute.manager [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: 6b19c327-9a80-498f-a350-8068ad250377] Took 1.03 seconds to deallocate network for instance. [ 724.044342] env[62600]: DEBUG nova.network.neutron [-] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.107311] env[62600]: DEBUG nova.compute.manager [req-3d6a1892-6e6b-4521-a519-7ef5347a9e0e req-ebfe01ea-bf1d-463c-80fc-b20ebd2a0479 service nova] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Received event network-changed-7714dcfd-c51e-4fd3-ae24-8fed6b64d011 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 724.107600] env[62600]: DEBUG nova.compute.manager [req-3d6a1892-6e6b-4521-a519-7ef5347a9e0e req-ebfe01ea-bf1d-463c-80fc-b20ebd2a0479 service nova] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Refreshing instance network info cache due to event network-changed-7714dcfd-c51e-4fd3-ae24-8fed6b64d011. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 724.107841] env[62600]: DEBUG oslo_concurrency.lockutils [req-3d6a1892-6e6b-4521-a519-7ef5347a9e0e req-ebfe01ea-bf1d-463c-80fc-b20ebd2a0479 service nova] Acquiring lock "refresh_cache-e3416076-1d11-47e4-9f76-9a64d7259166" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.107987] env[62600]: DEBUG oslo_concurrency.lockutils [req-3d6a1892-6e6b-4521-a519-7ef5347a9e0e req-ebfe01ea-bf1d-463c-80fc-b20ebd2a0479 service nova] Acquired lock "refresh_cache-e3416076-1d11-47e4-9f76-9a64d7259166" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.108403] env[62600]: DEBUG nova.network.neutron [req-3d6a1892-6e6b-4521-a519-7ef5347a9e0e req-ebfe01ea-bf1d-463c-80fc-b20ebd2a0479 service nova] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Refreshing network info cache for port 7714dcfd-c51e-4fd3-ae24-8fed6b64d011 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 724.220017] env[62600]: DEBUG oslo_concurrency.lockutils [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.482s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.220622] env[62600]: DEBUG nova.compute.manager [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 724.223662] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.159s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.225668] env[62600]: INFO nova.compute.claims [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 724.551225] env[62600]: INFO nova.compute.manager [-] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Took 1.02 seconds to deallocate network for instance. [ 724.553529] env[62600]: DEBUG nova.compute.claims [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 724.553529] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.630973] env[62600]: DEBUG nova.network.neutron [req-3d6a1892-6e6b-4521-a519-7ef5347a9e0e req-ebfe01ea-bf1d-463c-80fc-b20ebd2a0479 service nova] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.709636] env[62600]: DEBUG nova.network.neutron [req-3d6a1892-6e6b-4521-a519-7ef5347a9e0e req-ebfe01ea-bf1d-463c-80fc-b20ebd2a0479 service nova] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.731023] env[62600]: DEBUG nova.compute.utils [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 724.736784] env[62600]: DEBUG nova.compute.manager [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 724.738361] env[62600]: DEBUG nova.network.neutron [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 724.782641] env[62600]: DEBUG nova.policy [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '429aeb4ba9c941558de096e121a2c888', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '88e6070e46b64ae3b21719fb26c19345', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 724.972866] env[62600]: INFO nova.scheduler.client.report [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Deleted allocations for instance 6b19c327-9a80-498f-a350-8068ad250377 [ 725.073124] env[62600]: DEBUG nova.network.neutron [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Successfully created port: c24a86a9-b949-4ec9-b539-4c2f1158a89d {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 725.213166] env[62600]: DEBUG oslo_concurrency.lockutils [req-3d6a1892-6e6b-4521-a519-7ef5347a9e0e req-ebfe01ea-bf1d-463c-80fc-b20ebd2a0479 service nova] Releasing lock "refresh_cache-e3416076-1d11-47e4-9f76-9a64d7259166" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.213457] env[62600]: DEBUG nova.compute.manager [req-3d6a1892-6e6b-4521-a519-7ef5347a9e0e req-ebfe01ea-bf1d-463c-80fc-b20ebd2a0479 service nova] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Received event network-vif-deleted-7714dcfd-c51e-4fd3-ae24-8fed6b64d011 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 725.240348] env[62600]: DEBUG nova.compute.manager [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 725.485433] env[62600]: DEBUG oslo_concurrency.lockutils [None req-164787c2-dc0d-4791-a7e5-9298ea4f1f44 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "6b19c327-9a80-498f-a350-8068ad250377" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.306s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.711759] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce68e14d-c967-4924-86e2-4ab33e42012b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.720780] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be9044de-c5a6-4af9-955d-3c1a64a5c210 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.758835] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdd59c36-dfe5-409c-8cf9-3d6a4d75224b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.768258] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e83c16d4-8a4f-4a6c-92e7-18f097a2394e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.781924] env[62600]: DEBUG nova.compute.provider_tree [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.991721] env[62600]: DEBUG nova.compute.manager [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 726.144726] env[62600]: DEBUG nova.compute.manager [req-139f8a35-ab78-45c0-a639-0f0b4d21055b req-a662ad8a-3b8e-40d5-a6e6-ca9f0f16f2dc service nova] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Received event network-changed-c24a86a9-b949-4ec9-b539-4c2f1158a89d {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 726.144837] env[62600]: DEBUG nova.compute.manager [req-139f8a35-ab78-45c0-a639-0f0b4d21055b req-a662ad8a-3b8e-40d5-a6e6-ca9f0f16f2dc service nova] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Refreshing instance network info cache due to event network-changed-c24a86a9-b949-4ec9-b539-4c2f1158a89d. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 726.145102] env[62600]: DEBUG oslo_concurrency.lockutils [req-139f8a35-ab78-45c0-a639-0f0b4d21055b req-a662ad8a-3b8e-40d5-a6e6-ca9f0f16f2dc service nova] Acquiring lock "refresh_cache-877f5b67-68f8-48b0-8e12-622d9b50a390" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.145307] env[62600]: DEBUG oslo_concurrency.lockutils [req-139f8a35-ab78-45c0-a639-0f0b4d21055b req-a662ad8a-3b8e-40d5-a6e6-ca9f0f16f2dc service nova] Acquired lock "refresh_cache-877f5b67-68f8-48b0-8e12-622d9b50a390" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.145515] env[62600]: DEBUG nova.network.neutron [req-139f8a35-ab78-45c0-a639-0f0b4d21055b req-a662ad8a-3b8e-40d5-a6e6-ca9f0f16f2dc service nova] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Refreshing network info cache for port c24a86a9-b949-4ec9-b539-4c2f1158a89d {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 726.264527] env[62600]: DEBUG nova.compute.manager [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 726.285483] env[62600]: DEBUG nova.scheduler.client.report [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 726.306112] env[62600]: DEBUG nova.virt.hardware [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 726.306112] env[62600]: DEBUG nova.virt.hardware [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 726.306112] env[62600]: DEBUG nova.virt.hardware [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 726.306345] env[62600]: DEBUG nova.virt.hardware [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 726.306345] env[62600]: DEBUG nova.virt.hardware [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 726.306345] env[62600]: DEBUG nova.virt.hardware [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 726.306345] env[62600]: DEBUG nova.virt.hardware [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 726.306345] env[62600]: DEBUG nova.virt.hardware [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 726.306572] env[62600]: DEBUG nova.virt.hardware [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 726.306572] env[62600]: DEBUG nova.virt.hardware [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 726.306572] env[62600]: DEBUG nova.virt.hardware [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 726.306572] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2bd28ea-b5a2-482e-9fa8-f6ff5e25331a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.315443] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe5e9e0-75db-439d-9bd7-325560061509 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.383220] env[62600]: ERROR nova.compute.manager [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c24a86a9-b949-4ec9-b539-4c2f1158a89d, please check neutron logs for more information. [ 726.383220] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 726.383220] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.383220] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 726.383220] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 726.383220] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 726.383220] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 726.383220] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 726.383220] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.383220] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 726.383220] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.383220] env[62600]: ERROR nova.compute.manager raise self.value [ 726.383220] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 726.383220] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 726.383220] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.383220] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 726.383649] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.383649] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 726.383649] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c24a86a9-b949-4ec9-b539-4c2f1158a89d, please check neutron logs for more information. [ 726.383649] env[62600]: ERROR nova.compute.manager [ 726.383649] env[62600]: Traceback (most recent call last): [ 726.383649] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 726.383649] env[62600]: listener.cb(fileno) [ 726.383649] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.383649] env[62600]: result = function(*args, **kwargs) [ 726.383649] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.383649] env[62600]: return func(*args, **kwargs) [ 726.383649] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.383649] env[62600]: raise e [ 726.383649] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.383649] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 726.383649] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 726.383649] env[62600]: created_port_ids = self._update_ports_for_instance( [ 726.383649] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 726.383649] env[62600]: with excutils.save_and_reraise_exception(): [ 726.383649] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.383649] env[62600]: self.force_reraise() [ 726.383649] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.383649] env[62600]: raise self.value [ 726.383649] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 726.383649] env[62600]: updated_port = self._update_port( [ 726.383649] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.383649] env[62600]: _ensure_no_port_binding_failure(port) [ 726.383649] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.383649] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 726.384456] env[62600]: nova.exception.PortBindingFailed: Binding failed for port c24a86a9-b949-4ec9-b539-4c2f1158a89d, please check neutron logs for more information. [ 726.384456] env[62600]: Removing descriptor: 15 [ 726.384456] env[62600]: ERROR nova.compute.manager [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c24a86a9-b949-4ec9-b539-4c2f1158a89d, please check neutron logs for more information. [ 726.384456] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Traceback (most recent call last): [ 726.384456] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 726.384456] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] yield resources [ 726.384456] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 726.384456] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] self.driver.spawn(context, instance, image_meta, [ 726.384456] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 726.384456] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.384456] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.384456] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] vm_ref = self.build_virtual_machine(instance, [ 726.384797] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.384797] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.384797] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.384797] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] for vif in network_info: [ 726.384797] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 726.384797] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] return self._sync_wrapper(fn, *args, **kwargs) [ 726.384797] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 726.384797] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] self.wait() [ 726.384797] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 726.384797] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] self[:] = self._gt.wait() [ 726.384797] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.384797] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] return self._exit_event.wait() [ 726.384797] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 726.385192] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] result = hub.switch() [ 726.385192] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 726.385192] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] return self.greenlet.switch() [ 726.385192] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.385192] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] result = function(*args, **kwargs) [ 726.385192] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.385192] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] return func(*args, **kwargs) [ 726.385192] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.385192] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] raise e [ 726.385192] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.385192] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] nwinfo = self.network_api.allocate_for_instance( [ 726.385192] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 726.385192] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] created_port_ids = self._update_ports_for_instance( [ 726.385578] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 726.385578] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] with excutils.save_and_reraise_exception(): [ 726.385578] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.385578] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] self.force_reraise() [ 726.385578] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.385578] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] raise self.value [ 726.385578] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 726.385578] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] updated_port = self._update_port( [ 726.385578] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.385578] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] _ensure_no_port_binding_failure(port) [ 726.385578] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.385578] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] raise exception.PortBindingFailed(port_id=port['id']) [ 726.386267] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] nova.exception.PortBindingFailed: Binding failed for port c24a86a9-b949-4ec9-b539-4c2f1158a89d, please check neutron logs for more information. [ 726.386267] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] [ 726.386267] env[62600]: INFO nova.compute.manager [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Terminating instance [ 726.386615] env[62600]: DEBUG oslo_concurrency.lockutils [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Acquiring lock "refresh_cache-877f5b67-68f8-48b0-8e12-622d9b50a390" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.512704] env[62600]: DEBUG oslo_concurrency.lockutils [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.632801] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Acquiring lock "4e1a376f-6619-4c35-b75f-b45db75815cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.633047] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Lock "4e1a376f-6619-4c35-b75f-b45db75815cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.684314] env[62600]: DEBUG nova.network.neutron [req-139f8a35-ab78-45c0-a639-0f0b4d21055b req-a662ad8a-3b8e-40d5-a6e6-ca9f0f16f2dc service nova] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 726.791874] env[62600]: DEBUG nova.network.neutron [req-139f8a35-ab78-45c0-a639-0f0b4d21055b req-a662ad8a-3b8e-40d5-a6e6-ca9f0f16f2dc service nova] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.800632] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.573s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.800632] env[62600]: DEBUG nova.compute.manager [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 726.800632] env[62600]: DEBUG oslo_concurrency.lockutils [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.547s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.298610] env[62600]: DEBUG oslo_concurrency.lockutils [req-139f8a35-ab78-45c0-a639-0f0b4d21055b req-a662ad8a-3b8e-40d5-a6e6-ca9f0f16f2dc service nova] Releasing lock "refresh_cache-877f5b67-68f8-48b0-8e12-622d9b50a390" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.299029] env[62600]: DEBUG oslo_concurrency.lockutils [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Acquired lock "refresh_cache-877f5b67-68f8-48b0-8e12-622d9b50a390" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.299211] env[62600]: DEBUG nova.network.neutron [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 727.315023] env[62600]: DEBUG nova.compute.utils [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 727.319745] env[62600]: DEBUG nova.compute.manager [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 727.319745] env[62600]: DEBUG nova.network.neutron [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 727.366831] env[62600]: DEBUG nova.policy [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '46e18992b34d412e8e3e8205d1d18265', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '734eda84fbe4493ba1f3db243a2bdb09', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 727.642821] env[62600]: DEBUG nova.network.neutron [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Successfully created port: 9d61a4cd-49a3-4f42-8a41-7d5aecfa0db3 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 727.720496] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-918aef8d-d144-418f-bb61-5d57743a90ea {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.728073] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e84e09dd-0527-4eb0-be9d-da55360cf713 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.759935] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa8bcef8-2dbe-431b-8175-05293f3bc203 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.771023] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f500461f-1714-47f3-8bf3-5982fec3e1a9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.785587] env[62600]: DEBUG nova.compute.provider_tree [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.817819] env[62600]: DEBUG nova.compute.manager [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 727.826745] env[62600]: DEBUG nova.network.neutron [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 727.945393] env[62600]: DEBUG nova.network.neutron [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.170111] env[62600]: DEBUG nova.compute.manager [req-529d72e5-68d9-442d-8bd5-678a377e4be6 req-ccd9b4a7-40fa-4b49-a699-6bad578979d3 service nova] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Received event network-vif-deleted-c24a86a9-b949-4ec9-b539-4c2f1158a89d {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 728.288899] env[62600]: DEBUG nova.scheduler.client.report [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 728.360601] env[62600]: DEBUG oslo_concurrency.lockutils [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquiring lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.360869] env[62600]: DEBUG oslo_concurrency.lockutils [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.447918] env[62600]: DEBUG oslo_concurrency.lockutils [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Releasing lock "refresh_cache-877f5b67-68f8-48b0-8e12-622d9b50a390" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.448430] env[62600]: DEBUG nova.compute.manager [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 728.448632] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 728.449316] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f70dfbc2-f654-42e1-9ef1-1e013a4c530b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.462139] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c454ca2d-f5d3-4ae2-b1cd-a983bf768b80 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.485285] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 877f5b67-68f8-48b0-8e12-622d9b50a390 could not be found. [ 728.485544] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 728.485766] env[62600]: INFO nova.compute.manager [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Took 0.04 seconds to destroy the instance on the hypervisor. [ 728.486063] env[62600]: DEBUG oslo.service.loopingcall [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 728.486319] env[62600]: DEBUG nova.compute.manager [-] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 728.486422] env[62600]: DEBUG nova.network.neutron [-] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 728.506651] env[62600]: DEBUG nova.network.neutron [-] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 728.544955] env[62600]: ERROR nova.compute.manager [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9d61a4cd-49a3-4f42-8a41-7d5aecfa0db3, please check neutron logs for more information. [ 728.544955] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 728.544955] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.544955] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 728.544955] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 728.544955] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 728.544955] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 728.544955] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 728.544955] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.544955] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 728.544955] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.544955] env[62600]: ERROR nova.compute.manager raise self.value [ 728.544955] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 728.544955] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 728.544955] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.544955] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 728.545376] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.545376] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 728.545376] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9d61a4cd-49a3-4f42-8a41-7d5aecfa0db3, please check neutron logs for more information. [ 728.545376] env[62600]: ERROR nova.compute.manager [ 728.545376] env[62600]: Traceback (most recent call last): [ 728.545376] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 728.545376] env[62600]: listener.cb(fileno) [ 728.545376] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.545376] env[62600]: result = function(*args, **kwargs) [ 728.545376] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 728.545376] env[62600]: return func(*args, **kwargs) [ 728.545376] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.545376] env[62600]: raise e [ 728.545376] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.545376] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 728.545376] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 728.545376] env[62600]: created_port_ids = self._update_ports_for_instance( [ 728.545376] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 728.545376] env[62600]: with excutils.save_and_reraise_exception(): [ 728.545376] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.545376] env[62600]: self.force_reraise() [ 728.545376] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.545376] env[62600]: raise self.value [ 728.545376] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 728.545376] env[62600]: updated_port = self._update_port( [ 728.545376] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.545376] env[62600]: _ensure_no_port_binding_failure(port) [ 728.545376] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.545376] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 728.546078] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 9d61a4cd-49a3-4f42-8a41-7d5aecfa0db3, please check neutron logs for more information. [ 728.546078] env[62600]: Removing descriptor: 15 [ 728.795656] env[62600]: DEBUG oslo_concurrency.lockutils [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.996s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.796303] env[62600]: ERROR nova.compute.manager [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d0ae7b46-9488-45b2-89ed-a66a603cdec1, please check neutron logs for more information. [ 728.796303] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Traceback (most recent call last): [ 728.796303] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 728.796303] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] self.driver.spawn(context, instance, image_meta, [ 728.796303] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 728.796303] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.796303] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.796303] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] vm_ref = self.build_virtual_machine(instance, [ 728.796303] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.796303] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.796303] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.796647] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] for vif in network_info: [ 728.796647] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 728.796647] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] return self._sync_wrapper(fn, *args, **kwargs) [ 728.796647] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 728.796647] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] self.wait() [ 728.796647] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 728.796647] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] self[:] = self._gt.wait() [ 728.796647] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.796647] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] return self._exit_event.wait() [ 728.796647] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 728.796647] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] result = hub.switch() [ 728.796647] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 728.796647] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] return self.greenlet.switch() [ 728.797063] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.797063] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] result = function(*args, **kwargs) [ 728.797063] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 728.797063] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] return func(*args, **kwargs) [ 728.797063] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.797063] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] raise e [ 728.797063] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.797063] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] nwinfo = self.network_api.allocate_for_instance( [ 728.797063] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 728.797063] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] created_port_ids = self._update_ports_for_instance( [ 728.797063] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 728.797063] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] with excutils.save_and_reraise_exception(): [ 728.797063] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.797444] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] self.force_reraise() [ 728.797444] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.797444] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] raise self.value [ 728.797444] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 728.797444] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] updated_port = self._update_port( [ 728.797444] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.797444] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] _ensure_no_port_binding_failure(port) [ 728.797444] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.797444] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] raise exception.PortBindingFailed(port_id=port['id']) [ 728.797444] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] nova.exception.PortBindingFailed: Binding failed for port d0ae7b46-9488-45b2-89ed-a66a603cdec1, please check neutron logs for more information. [ 728.797444] env[62600]: ERROR nova.compute.manager [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] [ 728.798459] env[62600]: DEBUG nova.compute.utils [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Binding failed for port d0ae7b46-9488-45b2-89ed-a66a603cdec1, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 728.798459] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.193s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.799816] env[62600]: INFO nova.compute.claims [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 728.802336] env[62600]: DEBUG nova.compute.manager [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Build of instance 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b was re-scheduled: Binding failed for port d0ae7b46-9488-45b2-89ed-a66a603cdec1, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 728.802751] env[62600]: DEBUG nova.compute.manager [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 728.802971] env[62600]: DEBUG oslo_concurrency.lockutils [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Acquiring lock "refresh_cache-15772d8e-98ce-47cf-9b61-e79c7ffc3a8b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.803151] env[62600]: DEBUG oslo_concurrency.lockutils [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Acquired lock "refresh_cache-15772d8e-98ce-47cf-9b61-e79c7ffc3a8b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.803312] env[62600]: DEBUG nova.network.neutron [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 728.826649] env[62600]: DEBUG nova.compute.manager [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 728.852742] env[62600]: DEBUG nova.virt.hardware [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 728.852997] env[62600]: DEBUG nova.virt.hardware [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 728.853171] env[62600]: DEBUG nova.virt.hardware [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 728.853352] env[62600]: DEBUG nova.virt.hardware [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 728.853498] env[62600]: DEBUG nova.virt.hardware [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 728.853644] env[62600]: DEBUG nova.virt.hardware [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 728.853901] env[62600]: DEBUG nova.virt.hardware [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 728.854016] env[62600]: DEBUG nova.virt.hardware [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 728.854188] env[62600]: DEBUG nova.virt.hardware [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 728.854349] env[62600]: DEBUG nova.virt.hardware [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 728.854522] env[62600]: DEBUG nova.virt.hardware [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 728.855611] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a55886aa-468e-40c7-8ddf-7d250647a2ea {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.863343] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-583901f4-7bb6-4ab6-94a9-e99d0d4d7cb9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.877492] env[62600]: ERROR nova.compute.manager [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9d61a4cd-49a3-4f42-8a41-7d5aecfa0db3, please check neutron logs for more information. [ 728.877492] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Traceback (most recent call last): [ 728.877492] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 728.877492] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] yield resources [ 728.877492] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 728.877492] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] self.driver.spawn(context, instance, image_meta, [ 728.877492] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 728.877492] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.877492] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.877492] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] vm_ref = self.build_virtual_machine(instance, [ 728.877492] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.877829] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.877829] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.877829] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] for vif in network_info: [ 728.877829] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 728.877829] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] return self._sync_wrapper(fn, *args, **kwargs) [ 728.877829] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 728.877829] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] self.wait() [ 728.877829] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 728.877829] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] self[:] = self._gt.wait() [ 728.877829] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.877829] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] return self._exit_event.wait() [ 728.877829] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 728.877829] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] current.throw(*self._exc) [ 728.878180] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.878180] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] result = function(*args, **kwargs) [ 728.878180] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 728.878180] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] return func(*args, **kwargs) [ 728.878180] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.878180] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] raise e [ 728.878180] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.878180] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] nwinfo = self.network_api.allocate_for_instance( [ 728.878180] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 728.878180] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] created_port_ids = self._update_ports_for_instance( [ 728.878180] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 728.878180] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] with excutils.save_and_reraise_exception(): [ 728.878180] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.878566] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] self.force_reraise() [ 728.878566] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.878566] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] raise self.value [ 728.878566] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 728.878566] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] updated_port = self._update_port( [ 728.878566] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.878566] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] _ensure_no_port_binding_failure(port) [ 728.878566] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.878566] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] raise exception.PortBindingFailed(port_id=port['id']) [ 728.878566] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] nova.exception.PortBindingFailed: Binding failed for port 9d61a4cd-49a3-4f42-8a41-7d5aecfa0db3, please check neutron logs for more information. [ 728.878566] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] [ 728.878566] env[62600]: INFO nova.compute.manager [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Terminating instance [ 728.879752] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "refresh_cache-d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.879901] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquired lock "refresh_cache-d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.880093] env[62600]: DEBUG nova.network.neutron [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 729.008880] env[62600]: DEBUG nova.network.neutron [-] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.325805] env[62600]: DEBUG nova.network.neutron [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 729.406205] env[62600]: DEBUG nova.network.neutron [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 729.444426] env[62600]: DEBUG nova.network.neutron [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.496447] env[62600]: DEBUG nova.network.neutron [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.512149] env[62600]: INFO nova.compute.manager [-] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Took 1.03 seconds to deallocate network for instance. [ 729.514665] env[62600]: DEBUG nova.compute.claims [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 729.515568] env[62600]: DEBUG oslo_concurrency.lockutils [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.948718] env[62600]: DEBUG oslo_concurrency.lockutils [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Releasing lock "refresh_cache-15772d8e-98ce-47cf-9b61-e79c7ffc3a8b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.948968] env[62600]: DEBUG nova.compute.manager [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 729.949158] env[62600]: DEBUG nova.compute.manager [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 729.949737] env[62600]: DEBUG nova.network.neutron [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 729.966099] env[62600]: DEBUG nova.network.neutron [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 729.999386] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Releasing lock "refresh_cache-d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.999734] env[62600]: DEBUG nova.compute.manager [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 729.999920] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 730.000222] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e7fbf220-67bb-45a8-86f8-bcd0b46b559e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.010954] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5110167-9e21-4467-acea-6dbf81f9bae8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.034722] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d could not be found. [ 730.034958] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 730.035201] env[62600]: INFO nova.compute.manager [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 730.035447] env[62600]: DEBUG oslo.service.loopingcall [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 730.038021] env[62600]: DEBUG nova.compute.manager [-] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 730.038135] env[62600]: DEBUG nova.network.neutron [-] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 730.059381] env[62600]: DEBUG nova.network.neutron [-] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 730.145879] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b375c81b-79ed-43e1-a271-af6ddfcba82a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.153687] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9910df9e-4567-4423-a531-41f8639c4587 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.183521] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047d4c33-b6ce-48f0-a3df-1cd9fb1dd7b2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.192907] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-968cc808-d5e3-4bda-beb2-ae98a46ec024 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.197560] env[62600]: DEBUG nova.compute.manager [req-5421795f-d912-4e6b-9e83-89b83d6f6695 req-9212511c-ec64-44de-b522-c059291f0024 service nova] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Received event network-changed-9d61a4cd-49a3-4f42-8a41-7d5aecfa0db3 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 730.197745] env[62600]: DEBUG nova.compute.manager [req-5421795f-d912-4e6b-9e83-89b83d6f6695 req-9212511c-ec64-44de-b522-c059291f0024 service nova] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Refreshing instance network info cache due to event network-changed-9d61a4cd-49a3-4f42-8a41-7d5aecfa0db3. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 730.198091] env[62600]: DEBUG oslo_concurrency.lockutils [req-5421795f-d912-4e6b-9e83-89b83d6f6695 req-9212511c-ec64-44de-b522-c059291f0024 service nova] Acquiring lock "refresh_cache-d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.198197] env[62600]: DEBUG oslo_concurrency.lockutils [req-5421795f-d912-4e6b-9e83-89b83d6f6695 req-9212511c-ec64-44de-b522-c059291f0024 service nova] Acquired lock "refresh_cache-d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.198392] env[62600]: DEBUG nova.network.neutron [req-5421795f-d912-4e6b-9e83-89b83d6f6695 req-9212511c-ec64-44de-b522-c059291f0024 service nova] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Refreshing network info cache for port 9d61a4cd-49a3-4f42-8a41-7d5aecfa0db3 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 730.209487] env[62600]: DEBUG nova.compute.provider_tree [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.469029] env[62600]: DEBUG nova.network.neutron [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.563973] env[62600]: DEBUG nova.network.neutron [-] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.712512] env[62600]: DEBUG nova.scheduler.client.report [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 730.718499] env[62600]: DEBUG nova.network.neutron [req-5421795f-d912-4e6b-9e83-89b83d6f6695 req-9212511c-ec64-44de-b522-c059291f0024 service nova] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 730.789783] env[62600]: DEBUG nova.network.neutron [req-5421795f-d912-4e6b-9e83-89b83d6f6695 req-9212511c-ec64-44de-b522-c059291f0024 service nova] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.971861] env[62600]: INFO nova.compute.manager [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] [instance: 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b] Took 1.02 seconds to deallocate network for instance. [ 731.067843] env[62600]: INFO nova.compute.manager [-] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Took 1.03 seconds to deallocate network for instance. [ 731.068660] env[62600]: DEBUG nova.compute.claims [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 731.069149] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.217045] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.419s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.217579] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 731.220090] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.694s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.221465] env[62600]: INFO nova.compute.claims [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 731.291592] env[62600]: DEBUG oslo_concurrency.lockutils [req-5421795f-d912-4e6b-9e83-89b83d6f6695 req-9212511c-ec64-44de-b522-c059291f0024 service nova] Releasing lock "refresh_cache-d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.291913] env[62600]: DEBUG nova.compute.manager [req-5421795f-d912-4e6b-9e83-89b83d6f6695 req-9212511c-ec64-44de-b522-c059291f0024 service nova] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Received event network-vif-deleted-9d61a4cd-49a3-4f42-8a41-7d5aecfa0db3 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 731.726041] env[62600]: DEBUG nova.compute.utils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 731.729220] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 731.729406] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 731.768838] env[62600]: DEBUG nova.policy [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8b2c2a464fff4bef90406d02a72f8802', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2dcc8c2ea7b64be793ad62875b715b4e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 732.009140] env[62600]: INFO nova.scheduler.client.report [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Deleted allocations for instance 15772d8e-98ce-47cf-9b61-e79c7ffc3a8b [ 732.028959] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Successfully created port: 3388af72-15ab-44e3-ab44-8a8ad11950a9 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 732.230071] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 732.518356] env[62600]: DEBUG oslo_concurrency.lockutils [None req-955dfe43-7cbc-4c79-91e0-809d9e9c4faa tempest-ServerRescueTestJSONUnderV235-1856965101 tempest-ServerRescueTestJSONUnderV235-1856965101-project-member] Lock "15772d8e-98ce-47cf-9b61-e79c7ffc3a8b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.670s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.564500] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62fd4db0-4b0b-48d5-9351-82d6d63d2242 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.572149] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5fef500-5356-471a-9f01-0cad1a5cb3c3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.605427] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98d72cf3-1c82-4ba0-9c81-5b3f18b2aee2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.614019] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3434a801-01de-4946-931d-4d9a49a1388c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.627281] env[62600]: DEBUG nova.compute.provider_tree [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.705991] env[62600]: DEBUG nova.compute.manager [req-54979484-ca81-4b84-9292-b776e9bd1ad0 req-c789334d-db20-4b16-b2e1-1c5fce6a0796 service nova] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Received event network-changed-3388af72-15ab-44e3-ab44-8a8ad11950a9 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 732.706705] env[62600]: DEBUG nova.compute.manager [req-54979484-ca81-4b84-9292-b776e9bd1ad0 req-c789334d-db20-4b16-b2e1-1c5fce6a0796 service nova] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Refreshing instance network info cache due to event network-changed-3388af72-15ab-44e3-ab44-8a8ad11950a9. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 732.706885] env[62600]: DEBUG oslo_concurrency.lockutils [req-54979484-ca81-4b84-9292-b776e9bd1ad0 req-c789334d-db20-4b16-b2e1-1c5fce6a0796 service nova] Acquiring lock "refresh_cache-f989a4e4-c3c6-4cb5-9464-cbfb9d66c202" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.707100] env[62600]: DEBUG oslo_concurrency.lockutils [req-54979484-ca81-4b84-9292-b776e9bd1ad0 req-c789334d-db20-4b16-b2e1-1c5fce6a0796 service nova] Acquired lock "refresh_cache-f989a4e4-c3c6-4cb5-9464-cbfb9d66c202" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.707453] env[62600]: DEBUG nova.network.neutron [req-54979484-ca81-4b84-9292-b776e9bd1ad0 req-c789334d-db20-4b16-b2e1-1c5fce6a0796 service nova] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Refreshing network info cache for port 3388af72-15ab-44e3-ab44-8a8ad11950a9 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 732.914549] env[62600]: ERROR nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3388af72-15ab-44e3-ab44-8a8ad11950a9, please check neutron logs for more information. [ 732.914549] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 732.914549] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.914549] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 732.914549] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 732.914549] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 732.914549] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 732.914549] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 732.914549] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.914549] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 732.914549] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.914549] env[62600]: ERROR nova.compute.manager raise self.value [ 732.914549] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 732.914549] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 732.914549] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.914549] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 732.915345] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.915345] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 732.915345] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3388af72-15ab-44e3-ab44-8a8ad11950a9, please check neutron logs for more information. [ 732.915345] env[62600]: ERROR nova.compute.manager [ 732.915345] env[62600]: Traceback (most recent call last): [ 732.915345] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 732.915345] env[62600]: listener.cb(fileno) [ 732.915345] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.915345] env[62600]: result = function(*args, **kwargs) [ 732.915345] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 732.915345] env[62600]: return func(*args, **kwargs) [ 732.915345] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.915345] env[62600]: raise e [ 732.915345] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.915345] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 732.915345] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 732.915345] env[62600]: created_port_ids = self._update_ports_for_instance( [ 732.915345] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 732.915345] env[62600]: with excutils.save_and_reraise_exception(): [ 732.915345] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.915345] env[62600]: self.force_reraise() [ 732.915345] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.915345] env[62600]: raise self.value [ 732.915345] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 732.915345] env[62600]: updated_port = self._update_port( [ 732.915345] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.915345] env[62600]: _ensure_no_port_binding_failure(port) [ 732.915345] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.915345] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 732.916240] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 3388af72-15ab-44e3-ab44-8a8ad11950a9, please check neutron logs for more information. [ 732.916240] env[62600]: Removing descriptor: 16 [ 733.021872] env[62600]: DEBUG nova.compute.manager [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 733.131370] env[62600]: DEBUG nova.scheduler.client.report [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 733.241506] env[62600]: DEBUG nova.network.neutron [req-54979484-ca81-4b84-9292-b776e9bd1ad0 req-c789334d-db20-4b16-b2e1-1c5fce6a0796 service nova] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 733.246859] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 733.279228] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 733.279379] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 733.279428] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 733.279604] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 733.279743] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 733.279930] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 733.280160] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 733.280322] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 733.280483] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 733.280640] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 733.280831] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 733.281810] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ce51e35-b0a2-4de6-a79e-c4614945136b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.290267] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ae26b9-61bb-4194-a14e-9de690fe7957 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.305061] env[62600]: ERROR nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3388af72-15ab-44e3-ab44-8a8ad11950a9, please check neutron logs for more information. [ 733.305061] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Traceback (most recent call last): [ 733.305061] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 733.305061] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] yield resources [ 733.305061] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 733.305061] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] self.driver.spawn(context, instance, image_meta, [ 733.305061] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 733.305061] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] self._vmops.spawn(context, instance, image_meta, injected_files, [ 733.305061] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 733.305061] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] vm_ref = self.build_virtual_machine(instance, [ 733.305061] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 733.305511] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] vif_infos = vmwarevif.get_vif_info(self._session, [ 733.305511] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 733.305511] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] for vif in network_info: [ 733.305511] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 733.305511] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] return self._sync_wrapper(fn, *args, **kwargs) [ 733.305511] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 733.305511] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] self.wait() [ 733.305511] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 733.305511] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] self[:] = self._gt.wait() [ 733.305511] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 733.305511] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] return self._exit_event.wait() [ 733.305511] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 733.305511] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] current.throw(*self._exc) [ 733.305900] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.305900] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] result = function(*args, **kwargs) [ 733.305900] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 733.305900] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] return func(*args, **kwargs) [ 733.305900] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 733.305900] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] raise e [ 733.305900] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.305900] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] nwinfo = self.network_api.allocate_for_instance( [ 733.305900] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 733.305900] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] created_port_ids = self._update_ports_for_instance( [ 733.305900] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 733.305900] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] with excutils.save_and_reraise_exception(): [ 733.305900] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.306277] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] self.force_reraise() [ 733.306277] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.306277] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] raise self.value [ 733.306277] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 733.306277] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] updated_port = self._update_port( [ 733.306277] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.306277] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] _ensure_no_port_binding_failure(port) [ 733.306277] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.306277] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] raise exception.PortBindingFailed(port_id=port['id']) [ 733.306277] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] nova.exception.PortBindingFailed: Binding failed for port 3388af72-15ab-44e3-ab44-8a8ad11950a9, please check neutron logs for more information. [ 733.306277] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] [ 733.306277] env[62600]: INFO nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Terminating instance [ 733.309648] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquiring lock "refresh_cache-f989a4e4-c3c6-4cb5-9464-cbfb9d66c202" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.367747] env[62600]: DEBUG nova.network.neutron [req-54979484-ca81-4b84-9292-b776e9bd1ad0 req-c789334d-db20-4b16-b2e1-1c5fce6a0796 service nova] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.545875] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.637302] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.416s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.638664] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 733.646587] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.201s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.648245] env[62600]: INFO nova.compute.claims [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 733.869282] env[62600]: DEBUG oslo_concurrency.lockutils [req-54979484-ca81-4b84-9292-b776e9bd1ad0 req-c789334d-db20-4b16-b2e1-1c5fce6a0796 service nova] Releasing lock "refresh_cache-f989a4e4-c3c6-4cb5-9464-cbfb9d66c202" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.869742] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquired lock "refresh_cache-f989a4e4-c3c6-4cb5-9464-cbfb9d66c202" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.869930] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 734.153434] env[62600]: DEBUG nova.compute.utils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 734.157790] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 734.157995] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 734.211415] env[62600]: DEBUG nova.policy [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8b2c2a464fff4bef90406d02a72f8802', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2dcc8c2ea7b64be793ad62875b715b4e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 734.391139] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 734.510033] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.552113] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Successfully created port: e8122e39-0982-4178-97a6-52a3f6e5c9e8 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 734.657736] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 734.792066] env[62600]: DEBUG nova.compute.manager [req-5408749d-82d4-42ee-834e-852a701c5e10 req-7349a6ee-7aa0-4144-8642-630ce5043b3c service nova] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Received event network-vif-deleted-3388af72-15ab-44e3-ab44-8a8ad11950a9 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 735.014596] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Releasing lock "refresh_cache-f989a4e4-c3c6-4cb5-9464-cbfb9d66c202" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.015132] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 735.015342] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 735.015716] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2d9afa6a-5b2f-4844-9934-48123c1bd21f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.027398] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-560d5b9d-cbf2-473f-9f72-b31723b945b5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.054620] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f989a4e4-c3c6-4cb5-9464-cbfb9d66c202 could not be found. [ 735.054834] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 735.055160] env[62600]: INFO nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Took 0.04 seconds to destroy the instance on the hypervisor. [ 735.055352] env[62600]: DEBUG oslo.service.loopingcall [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 735.055540] env[62600]: DEBUG nova.compute.manager [-] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 735.055685] env[62600]: DEBUG nova.network.neutron [-] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 735.077570] env[62600]: DEBUG nova.network.neutron [-] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 735.094688] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7342888f-8259-46f4-85c7-f2526f803ba9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.103161] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b801e2d7-ea1e-4408-a456-5f49e67e9155 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.140195] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb6966f-97bb-4b01-9839-197dc5797553 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.147792] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b0d98c9-4be9-4466-8af7-77d0ec11b5b7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.161499] env[62600]: DEBUG nova.compute.provider_tree [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.514651] env[62600]: ERROR nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e8122e39-0982-4178-97a6-52a3f6e5c9e8, please check neutron logs for more information. [ 735.514651] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 735.514651] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.514651] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 735.514651] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 735.514651] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 735.514651] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 735.514651] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 735.514651] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.514651] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 735.514651] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.514651] env[62600]: ERROR nova.compute.manager raise self.value [ 735.514651] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 735.514651] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 735.514651] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.514651] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 735.515216] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.515216] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 735.515216] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e8122e39-0982-4178-97a6-52a3f6e5c9e8, please check neutron logs for more information. [ 735.515216] env[62600]: ERROR nova.compute.manager [ 735.515216] env[62600]: Traceback (most recent call last): [ 735.515216] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 735.515216] env[62600]: listener.cb(fileno) [ 735.515216] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.515216] env[62600]: result = function(*args, **kwargs) [ 735.515216] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 735.515216] env[62600]: return func(*args, **kwargs) [ 735.515216] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.515216] env[62600]: raise e [ 735.515216] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.515216] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 735.515216] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 735.515216] env[62600]: created_port_ids = self._update_ports_for_instance( [ 735.515216] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 735.515216] env[62600]: with excutils.save_and_reraise_exception(): [ 735.515216] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.515216] env[62600]: self.force_reraise() [ 735.515216] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.515216] env[62600]: raise self.value [ 735.515216] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 735.515216] env[62600]: updated_port = self._update_port( [ 735.515216] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.515216] env[62600]: _ensure_no_port_binding_failure(port) [ 735.515216] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.515216] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 735.515937] env[62600]: nova.exception.PortBindingFailed: Binding failed for port e8122e39-0982-4178-97a6-52a3f6e5c9e8, please check neutron logs for more information. [ 735.515937] env[62600]: Removing descriptor: 16 [ 735.580894] env[62600]: DEBUG nova.network.neutron [-] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.663887] env[62600]: DEBUG nova.scheduler.client.report [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 735.675304] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 735.701176] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 735.701416] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 735.701572] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.701753] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 735.701897] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.702054] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 735.702268] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 735.702428] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 735.702594] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 735.702755] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 735.702924] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 735.703781] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6daa5bea-07ec-41d1-8412-3d6962aa9ec6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.711430] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df8d19ff-5b04-4ef3-bc9c-b067df2752e9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.725914] env[62600]: ERROR nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e8122e39-0982-4178-97a6-52a3f6e5c9e8, please check neutron logs for more information. [ 735.725914] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Traceback (most recent call last): [ 735.725914] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 735.725914] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] yield resources [ 735.725914] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 735.725914] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] self.driver.spawn(context, instance, image_meta, [ 735.725914] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 735.725914] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 735.725914] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 735.725914] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] vm_ref = self.build_virtual_machine(instance, [ 735.725914] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 735.726433] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] vif_infos = vmwarevif.get_vif_info(self._session, [ 735.726433] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 735.726433] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] for vif in network_info: [ 735.726433] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 735.726433] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] return self._sync_wrapper(fn, *args, **kwargs) [ 735.726433] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 735.726433] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] self.wait() [ 735.726433] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 735.726433] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] self[:] = self._gt.wait() [ 735.726433] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 735.726433] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] return self._exit_event.wait() [ 735.726433] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 735.726433] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] current.throw(*self._exc) [ 735.726962] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.726962] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] result = function(*args, **kwargs) [ 735.726962] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 735.726962] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] return func(*args, **kwargs) [ 735.726962] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.726962] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] raise e [ 735.726962] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.726962] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] nwinfo = self.network_api.allocate_for_instance( [ 735.726962] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 735.726962] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] created_port_ids = self._update_ports_for_instance( [ 735.726962] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 735.726962] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] with excutils.save_and_reraise_exception(): [ 735.726962] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.727372] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] self.force_reraise() [ 735.727372] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.727372] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] raise self.value [ 735.727372] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 735.727372] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] updated_port = self._update_port( [ 735.727372] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.727372] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] _ensure_no_port_binding_failure(port) [ 735.727372] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.727372] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] raise exception.PortBindingFailed(port_id=port['id']) [ 735.727372] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] nova.exception.PortBindingFailed: Binding failed for port e8122e39-0982-4178-97a6-52a3f6e5c9e8, please check neutron logs for more information. [ 735.727372] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] [ 735.727372] env[62600]: INFO nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Terminating instance [ 735.727706] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquiring lock "refresh_cache-9c2c8dd4-85c7-489a-95c0-669644237b8b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.727706] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquired lock "refresh_cache-9c2c8dd4-85c7-489a-95c0-669644237b8b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.727706] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 736.084952] env[62600]: INFO nova.compute.manager [-] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Took 1.03 seconds to deallocate network for instance. [ 736.086021] env[62600]: DEBUG nova.compute.claims [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 736.086021] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.172025] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.172025] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 736.180015] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.707s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.180015] env[62600]: INFO nova.compute.claims [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 736.248283] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 736.330285] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.690985] env[62600]: DEBUG nova.compute.utils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 736.692485] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 736.692485] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 736.739365] env[62600]: DEBUG nova.policy [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8b2c2a464fff4bef90406d02a72f8802', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2dcc8c2ea7b64be793ad62875b715b4e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 736.830524] env[62600]: DEBUG nova.compute.manager [req-d2e0bc6c-7f1d-4747-b4d5-9c9d2c867d1e req-52afb2cd-c839-4acc-af7a-a5095315a5dc service nova] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Received event network-changed-e8122e39-0982-4178-97a6-52a3f6e5c9e8 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 736.830661] env[62600]: DEBUG nova.compute.manager [req-d2e0bc6c-7f1d-4747-b4d5-9c9d2c867d1e req-52afb2cd-c839-4acc-af7a-a5095315a5dc service nova] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Refreshing instance network info cache due to event network-changed-e8122e39-0982-4178-97a6-52a3f6e5c9e8. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 736.831474] env[62600]: DEBUG oslo_concurrency.lockutils [req-d2e0bc6c-7f1d-4747-b4d5-9c9d2c867d1e req-52afb2cd-c839-4acc-af7a-a5095315a5dc service nova] Acquiring lock "refresh_cache-9c2c8dd4-85c7-489a-95c0-669644237b8b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.832938] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Releasing lock "refresh_cache-9c2c8dd4-85c7-489a-95c0-669644237b8b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.833414] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 736.834383] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 736.834383] env[62600]: DEBUG oslo_concurrency.lockutils [req-d2e0bc6c-7f1d-4747-b4d5-9c9d2c867d1e req-52afb2cd-c839-4acc-af7a-a5095315a5dc service nova] Acquired lock "refresh_cache-9c2c8dd4-85c7-489a-95c0-669644237b8b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.834383] env[62600]: DEBUG nova.network.neutron [req-d2e0bc6c-7f1d-4747-b4d5-9c9d2c867d1e req-52afb2cd-c839-4acc-af7a-a5095315a5dc service nova] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Refreshing network info cache for port e8122e39-0982-4178-97a6-52a3f6e5c9e8 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 736.835123] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-89bcae51-414b-4a79-91eb-0004ea92ad18 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.849159] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-678b681f-6e37-486f-8607-3c30f7d4974d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.877627] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9c2c8dd4-85c7-489a-95c0-669644237b8b could not be found. [ 736.877747] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 736.877923] env[62600]: INFO nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 736.878187] env[62600]: DEBUG oslo.service.loopingcall [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 736.879035] env[62600]: DEBUG nova.compute.manager [-] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 736.879035] env[62600]: DEBUG nova.network.neutron [-] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 736.898397] env[62600]: DEBUG nova.network.neutron [-] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.039772] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Successfully created port: 47c7b2fe-dec4-439d-b0a7-f1b14dbf1f05 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 737.199859] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 737.362018] env[62600]: DEBUG nova.network.neutron [req-d2e0bc6c-7f1d-4747-b4d5-9c9d2c867d1e req-52afb2cd-c839-4acc-af7a-a5095315a5dc service nova] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.402727] env[62600]: DEBUG nova.network.neutron [-] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.508933] env[62600]: DEBUG nova.network.neutron [req-d2e0bc6c-7f1d-4747-b4d5-9c9d2c867d1e req-52afb2cd-c839-4acc-af7a-a5095315a5dc service nova] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.661016] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd131e6-32ba-43c1-89c9-98f7db737461 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.671018] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c03f1f9e-7d38-48e1-ab40-040dc9735058 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.701196] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df48033a-4610-4dbe-beb4-6f6c51635a77 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.711442] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bb65050-7dd3-4747-8295-75851fb13c39 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.725266] env[62600]: DEBUG nova.compute.provider_tree [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.905524] env[62600]: INFO nova.compute.manager [-] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Took 1.03 seconds to deallocate network for instance. [ 737.908172] env[62600]: DEBUG nova.compute.claims [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 737.908430] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.013470] env[62600]: DEBUG oslo_concurrency.lockutils [req-d2e0bc6c-7f1d-4747-b4d5-9c9d2c867d1e req-52afb2cd-c839-4acc-af7a-a5095315a5dc service nova] Releasing lock "refresh_cache-9c2c8dd4-85c7-489a-95c0-669644237b8b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.013758] env[62600]: DEBUG nova.compute.manager [req-d2e0bc6c-7f1d-4747-b4d5-9c9d2c867d1e req-52afb2cd-c839-4acc-af7a-a5095315a5dc service nova] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Received event network-vif-deleted-e8122e39-0982-4178-97a6-52a3f6e5c9e8 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 738.213082] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 738.229198] env[62600]: DEBUG nova.scheduler.client.report [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 738.257339] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 738.258212] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 738.258212] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 738.258212] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 738.258212] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 738.258409] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 738.258683] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 738.259057] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 738.259124] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 738.259341] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 738.259553] env[62600]: DEBUG nova.virt.hardware [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 738.260732] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f746eb80-60fa-4e1e-ae61-937ddb63df12 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.272403] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c32d7b1-8a65-4580-a201-9d4882f4ab04 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.278461] env[62600]: ERROR nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 47c7b2fe-dec4-439d-b0a7-f1b14dbf1f05, please check neutron logs for more information. [ 738.278461] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 738.278461] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.278461] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 738.278461] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 738.278461] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 738.278461] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 738.278461] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 738.278461] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.278461] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 738.278461] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.278461] env[62600]: ERROR nova.compute.manager raise self.value [ 738.278461] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 738.278461] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 738.278461] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.278461] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 738.278991] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.278991] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 738.278991] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 47c7b2fe-dec4-439d-b0a7-f1b14dbf1f05, please check neutron logs for more information. [ 738.278991] env[62600]: ERROR nova.compute.manager [ 738.278991] env[62600]: Traceback (most recent call last): [ 738.278991] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 738.278991] env[62600]: listener.cb(fileno) [ 738.278991] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.278991] env[62600]: result = function(*args, **kwargs) [ 738.278991] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 738.278991] env[62600]: return func(*args, **kwargs) [ 738.278991] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.278991] env[62600]: raise e [ 738.278991] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.278991] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 738.278991] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 738.278991] env[62600]: created_port_ids = self._update_ports_for_instance( [ 738.278991] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 738.278991] env[62600]: with excutils.save_and_reraise_exception(): [ 738.278991] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.278991] env[62600]: self.force_reraise() [ 738.278991] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.278991] env[62600]: raise self.value [ 738.278991] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 738.278991] env[62600]: updated_port = self._update_port( [ 738.278991] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.278991] env[62600]: _ensure_no_port_binding_failure(port) [ 738.278991] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.278991] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 738.279812] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 47c7b2fe-dec4-439d-b0a7-f1b14dbf1f05, please check neutron logs for more information. [ 738.279812] env[62600]: Removing descriptor: 16 [ 738.291315] env[62600]: ERROR nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 47c7b2fe-dec4-439d-b0a7-f1b14dbf1f05, please check neutron logs for more information. [ 738.291315] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] Traceback (most recent call last): [ 738.291315] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 738.291315] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] yield resources [ 738.291315] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 738.291315] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] self.driver.spawn(context, instance, image_meta, [ 738.291315] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 738.291315] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.291315] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.291315] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] vm_ref = self.build_virtual_machine(instance, [ 738.291315] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.291703] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.291703] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.291703] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] for vif in network_info: [ 738.291703] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 738.291703] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] return self._sync_wrapper(fn, *args, **kwargs) [ 738.291703] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 738.291703] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] self.wait() [ 738.291703] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 738.291703] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] self[:] = self._gt.wait() [ 738.291703] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.291703] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] return self._exit_event.wait() [ 738.291703] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 738.291703] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] current.throw(*self._exc) [ 738.292513] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.292513] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] result = function(*args, **kwargs) [ 738.292513] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 738.292513] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] return func(*args, **kwargs) [ 738.292513] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.292513] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] raise e [ 738.292513] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.292513] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] nwinfo = self.network_api.allocate_for_instance( [ 738.292513] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 738.292513] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] created_port_ids = self._update_ports_for_instance( [ 738.292513] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 738.292513] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] with excutils.save_and_reraise_exception(): [ 738.292513] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.292929] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] self.force_reraise() [ 738.292929] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.292929] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] raise self.value [ 738.292929] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 738.292929] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] updated_port = self._update_port( [ 738.292929] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.292929] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] _ensure_no_port_binding_failure(port) [ 738.292929] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.292929] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] raise exception.PortBindingFailed(port_id=port['id']) [ 738.292929] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] nova.exception.PortBindingFailed: Binding failed for port 47c7b2fe-dec4-439d-b0a7-f1b14dbf1f05, please check neutron logs for more information. [ 738.292929] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] [ 738.294029] env[62600]: INFO nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Terminating instance [ 738.296078] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquiring lock "refresh_cache-81814471-a617-4364-a140-105252ca9c04" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.296241] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquired lock "refresh_cache-81814471-a617-4364-a140-105252ca9c04" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.296422] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 738.736799] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.561s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.737809] env[62600]: DEBUG nova.compute.manager [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 738.741927] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.323s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.823451] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 738.868019] env[62600]: DEBUG nova.compute.manager [req-d00a56b1-88e0-4cf3-be17-159f92c27fed req-34a8ee9c-01ba-4b17-b359-8f49c30904d4 service nova] [instance: 81814471-a617-4364-a140-105252ca9c04] Received event network-changed-47c7b2fe-dec4-439d-b0a7-f1b14dbf1f05 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 738.868384] env[62600]: DEBUG nova.compute.manager [req-d00a56b1-88e0-4cf3-be17-159f92c27fed req-34a8ee9c-01ba-4b17-b359-8f49c30904d4 service nova] [instance: 81814471-a617-4364-a140-105252ca9c04] Refreshing instance network info cache due to event network-changed-47c7b2fe-dec4-439d-b0a7-f1b14dbf1f05. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 738.868458] env[62600]: DEBUG oslo_concurrency.lockutils [req-d00a56b1-88e0-4cf3-be17-159f92c27fed req-34a8ee9c-01ba-4b17-b359-8f49c30904d4 service nova] Acquiring lock "refresh_cache-81814471-a617-4364-a140-105252ca9c04" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.949890] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.245067] env[62600]: DEBUG nova.compute.utils [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 739.245367] env[62600]: DEBUG nova.compute.manager [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 739.246406] env[62600]: DEBUG nova.network.neutron [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 739.324229] env[62600]: DEBUG nova.policy [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7627870c06e14be4aedbd4ee9bee374f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8447a3f29a8347a0b3be942f5bdbc126', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 739.459246] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Releasing lock "refresh_cache-81814471-a617-4364-a140-105252ca9c04" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.459246] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 739.459246] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 739.459451] env[62600]: DEBUG oslo_concurrency.lockutils [req-d00a56b1-88e0-4cf3-be17-159f92c27fed req-34a8ee9c-01ba-4b17-b359-8f49c30904d4 service nova] Acquired lock "refresh_cache-81814471-a617-4364-a140-105252ca9c04" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.459611] env[62600]: DEBUG nova.network.neutron [req-d00a56b1-88e0-4cf3-be17-159f92c27fed req-34a8ee9c-01ba-4b17-b359-8f49c30904d4 service nova] [instance: 81814471-a617-4364-a140-105252ca9c04] Refreshing network info cache for port 47c7b2fe-dec4-439d-b0a7-f1b14dbf1f05 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 739.460782] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf6277ea-c81c-43d0-b37d-f386e34bf6ce {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.477854] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9310cceb-e5c6-4755-8c48-ab1570cf4012 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.502046] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 81814471-a617-4364-a140-105252ca9c04 could not be found. [ 739.502275] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 739.502449] env[62600]: INFO nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Took 0.04 seconds to destroy the instance on the hypervisor. [ 739.502712] env[62600]: DEBUG oslo.service.loopingcall [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 739.505214] env[62600]: DEBUG nova.compute.manager [-] [instance: 81814471-a617-4364-a140-105252ca9c04] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 739.505299] env[62600]: DEBUG nova.network.neutron [-] [instance: 81814471-a617-4364-a140-105252ca9c04] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 739.533093] env[62600]: DEBUG nova.network.neutron [-] [instance: 81814471-a617-4364-a140-105252ca9c04] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 739.748710] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b24bc7f-83f2-4d68-bcba-869b878e3a24 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.752805] env[62600]: DEBUG nova.compute.manager [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 739.759790] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f4c2fc6-cb78-42f8-b107-31d24808cc90 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.795845] env[62600]: DEBUG nova.network.neutron [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Successfully created port: 2f5a8c89-4486-4877-bb39-c7a87408ecc4 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 739.798221] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3939064-6ac3-4246-9c28-6b44354c88e9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.807214] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b337806-b933-4ed2-ad4b-0eab8505b4ef {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.822128] env[62600]: DEBUG nova.compute.provider_tree [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.993633] env[62600]: DEBUG nova.network.neutron [req-d00a56b1-88e0-4cf3-be17-159f92c27fed req-34a8ee9c-01ba-4b17-b359-8f49c30904d4 service nova] [instance: 81814471-a617-4364-a140-105252ca9c04] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 740.035857] env[62600]: DEBUG nova.network.neutron [-] [instance: 81814471-a617-4364-a140-105252ca9c04] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.122808] env[62600]: DEBUG nova.network.neutron [req-d00a56b1-88e0-4cf3-be17-159f92c27fed req-34a8ee9c-01ba-4b17-b359-8f49c30904d4 service nova] [instance: 81814471-a617-4364-a140-105252ca9c04] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.169896] env[62600]: DEBUG nova.network.neutron [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Successfully created port: 3fdbac1b-aa04-4705-9c77-40f4d52244e1 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 740.325035] env[62600]: DEBUG nova.scheduler.client.report [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 740.540362] env[62600]: INFO nova.compute.manager [-] [instance: 81814471-a617-4364-a140-105252ca9c04] Took 1.03 seconds to deallocate network for instance. [ 740.543573] env[62600]: DEBUG nova.compute.claims [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 740.543573] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.626849] env[62600]: DEBUG oslo_concurrency.lockutils [req-d00a56b1-88e0-4cf3-be17-159f92c27fed req-34a8ee9c-01ba-4b17-b359-8f49c30904d4 service nova] Releasing lock "refresh_cache-81814471-a617-4364-a140-105252ca9c04" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.627136] env[62600]: DEBUG nova.compute.manager [req-d00a56b1-88e0-4cf3-be17-159f92c27fed req-34a8ee9c-01ba-4b17-b359-8f49c30904d4 service nova] [instance: 81814471-a617-4364-a140-105252ca9c04] Received event network-vif-deleted-47c7b2fe-dec4-439d-b0a7-f1b14dbf1f05 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 740.762781] env[62600]: DEBUG nova.compute.manager [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 740.803747] env[62600]: DEBUG nova.virt.hardware [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 740.803955] env[62600]: DEBUG nova.virt.hardware [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 740.804114] env[62600]: DEBUG nova.virt.hardware [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 740.804620] env[62600]: DEBUG nova.virt.hardware [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 740.804724] env[62600]: DEBUG nova.virt.hardware [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 740.804972] env[62600]: DEBUG nova.virt.hardware [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 740.805085] env[62600]: DEBUG nova.virt.hardware [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 740.805287] env[62600]: DEBUG nova.virt.hardware [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 740.805465] env[62600]: DEBUG nova.virt.hardware [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 740.805631] env[62600]: DEBUG nova.virt.hardware [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 740.805803] env[62600]: DEBUG nova.virt.hardware [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 740.806988] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d9969f8-81d0-40d4-aa23-727729a7f2a3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.817538] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e4c1f9-6c45-431d-a5b4-5ea33ee567a5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.832202] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.090s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.832932] env[62600]: ERROR nova.compute.manager [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ad089818-0871-43ff-8fd3-58654980f370, please check neutron logs for more information. [ 740.832932] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] Traceback (most recent call last): [ 740.832932] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 740.832932] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] self.driver.spawn(context, instance, image_meta, [ 740.832932] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 740.832932] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] self._vmops.spawn(context, instance, image_meta, injected_files, [ 740.832932] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 740.832932] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] vm_ref = self.build_virtual_machine(instance, [ 740.832932] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 740.832932] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] vif_infos = vmwarevif.get_vif_info(self._session, [ 740.832932] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 740.833444] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] for vif in network_info: [ 740.833444] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 740.833444] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] return self._sync_wrapper(fn, *args, **kwargs) [ 740.833444] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 740.833444] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] self.wait() [ 740.833444] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 740.833444] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] self[:] = self._gt.wait() [ 740.833444] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 740.833444] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] return self._exit_event.wait() [ 740.833444] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 740.833444] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] result = hub.switch() [ 740.833444] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 740.833444] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] return self.greenlet.switch() [ 740.833945] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.833945] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] result = function(*args, **kwargs) [ 740.833945] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 740.833945] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] return func(*args, **kwargs) [ 740.833945] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 740.833945] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] raise e [ 740.833945] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.833945] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] nwinfo = self.network_api.allocate_for_instance( [ 740.833945] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 740.833945] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] created_port_ids = self._update_ports_for_instance( [ 740.833945] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 740.833945] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] with excutils.save_and_reraise_exception(): [ 740.833945] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.834327] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] self.force_reraise() [ 740.834327] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.834327] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] raise self.value [ 740.834327] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 740.834327] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] updated_port = self._update_port( [ 740.834327] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.834327] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] _ensure_no_port_binding_failure(port) [ 740.834327] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.834327] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] raise exception.PortBindingFailed(port_id=port['id']) [ 740.834327] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] nova.exception.PortBindingFailed: Binding failed for port ad089818-0871-43ff-8fd3-58654980f370, please check neutron logs for more information. [ 740.834327] env[62600]: ERROR nova.compute.manager [instance: c93cb228-0adb-423c-854f-73354cdb7012] [ 740.834634] env[62600]: DEBUG nova.compute.utils [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Binding failed for port ad089818-0871-43ff-8fd3-58654980f370, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 740.835890] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.282s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.839196] env[62600]: DEBUG nova.compute.manager [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Build of instance c93cb228-0adb-423c-854f-73354cdb7012 was re-scheduled: Binding failed for port ad089818-0871-43ff-8fd3-58654980f370, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 740.839851] env[62600]: DEBUG nova.compute.manager [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 740.839851] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "refresh_cache-c93cb228-0adb-423c-854f-73354cdb7012" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.840095] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired lock "refresh_cache-c93cb228-0adb-423c-854f-73354cdb7012" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.840309] env[62600]: DEBUG nova.network.neutron [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 740.936304] env[62600]: DEBUG nova.compute.manager [req-c71a984d-569f-4806-a85d-d9fe748594f3 req-da341b1c-dd18-4f57-877b-f811e811b181 service nova] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Received event network-changed-2f5a8c89-4486-4877-bb39-c7a87408ecc4 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 740.936304] env[62600]: DEBUG nova.compute.manager [req-c71a984d-569f-4806-a85d-d9fe748594f3 req-da341b1c-dd18-4f57-877b-f811e811b181 service nova] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Refreshing instance network info cache due to event network-changed-2f5a8c89-4486-4877-bb39-c7a87408ecc4. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 740.936304] env[62600]: DEBUG oslo_concurrency.lockutils [req-c71a984d-569f-4806-a85d-d9fe748594f3 req-da341b1c-dd18-4f57-877b-f811e811b181 service nova] Acquiring lock "refresh_cache-68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.936304] env[62600]: DEBUG oslo_concurrency.lockutils [req-c71a984d-569f-4806-a85d-d9fe748594f3 req-da341b1c-dd18-4f57-877b-f811e811b181 service nova] Acquired lock "refresh_cache-68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.936846] env[62600]: DEBUG nova.network.neutron [req-c71a984d-569f-4806-a85d-d9fe748594f3 req-da341b1c-dd18-4f57-877b-f811e811b181 service nova] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Refreshing network info cache for port 2f5a8c89-4486-4877-bb39-c7a87408ecc4 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 741.040172] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Acquiring lock "fb6f8e36-8d24-45ea-a6e4-4d768c3b232d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.043024] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Lock "fb6f8e36-8d24-45ea-a6e4-4d768c3b232d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.124414] env[62600]: ERROR nova.compute.manager [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2f5a8c89-4486-4877-bb39-c7a87408ecc4, please check neutron logs for more information. [ 741.124414] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 741.124414] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.124414] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 741.124414] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 741.124414] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 741.124414] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 741.124414] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 741.124414] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.124414] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 741.124414] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.124414] env[62600]: ERROR nova.compute.manager raise self.value [ 741.124414] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 741.124414] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 741.124414] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.124414] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 741.125246] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.125246] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 741.125246] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2f5a8c89-4486-4877-bb39-c7a87408ecc4, please check neutron logs for more information. [ 741.125246] env[62600]: ERROR nova.compute.manager [ 741.125246] env[62600]: Traceback (most recent call last): [ 741.125246] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 741.125246] env[62600]: listener.cb(fileno) [ 741.125246] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 741.125246] env[62600]: result = function(*args, **kwargs) [ 741.125246] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 741.125246] env[62600]: return func(*args, **kwargs) [ 741.125246] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 741.125246] env[62600]: raise e [ 741.125246] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.125246] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 741.125246] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 741.125246] env[62600]: created_port_ids = self._update_ports_for_instance( [ 741.125246] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 741.125246] env[62600]: with excutils.save_and_reraise_exception(): [ 741.125246] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.125246] env[62600]: self.force_reraise() [ 741.125246] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.125246] env[62600]: raise self.value [ 741.125246] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 741.125246] env[62600]: updated_port = self._update_port( [ 741.125246] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.125246] env[62600]: _ensure_no_port_binding_failure(port) [ 741.125246] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.125246] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 741.126242] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 2f5a8c89-4486-4877-bb39-c7a87408ecc4, please check neutron logs for more information. [ 741.126242] env[62600]: Removing descriptor: 16 [ 741.126242] env[62600]: ERROR nova.compute.manager [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2f5a8c89-4486-4877-bb39-c7a87408ecc4, please check neutron logs for more information. [ 741.126242] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Traceback (most recent call last): [ 741.126242] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 741.126242] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] yield resources [ 741.126242] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 741.126242] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] self.driver.spawn(context, instance, image_meta, [ 741.126242] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 741.126242] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 741.126242] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 741.126242] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] vm_ref = self.build_virtual_machine(instance, [ 741.126559] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 741.126559] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] vif_infos = vmwarevif.get_vif_info(self._session, [ 741.126559] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 741.126559] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] for vif in network_info: [ 741.126559] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 741.126559] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] return self._sync_wrapper(fn, *args, **kwargs) [ 741.126559] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 741.126559] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] self.wait() [ 741.126559] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 741.126559] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] self[:] = self._gt.wait() [ 741.126559] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 741.126559] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] return self._exit_event.wait() [ 741.126559] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 741.126905] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] result = hub.switch() [ 741.126905] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 741.126905] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] return self.greenlet.switch() [ 741.126905] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 741.126905] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] result = function(*args, **kwargs) [ 741.126905] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 741.126905] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] return func(*args, **kwargs) [ 741.126905] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 741.126905] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] raise e [ 741.126905] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.126905] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] nwinfo = self.network_api.allocate_for_instance( [ 741.126905] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 741.126905] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] created_port_ids = self._update_ports_for_instance( [ 741.127264] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 741.127264] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] with excutils.save_and_reraise_exception(): [ 741.127264] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.127264] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] self.force_reraise() [ 741.127264] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.127264] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] raise self.value [ 741.127264] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 741.127264] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] updated_port = self._update_port( [ 741.127264] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.127264] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] _ensure_no_port_binding_failure(port) [ 741.127264] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.127264] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] raise exception.PortBindingFailed(port_id=port['id']) [ 741.127579] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] nova.exception.PortBindingFailed: Binding failed for port 2f5a8c89-4486-4877-bb39-c7a87408ecc4, please check neutron logs for more information. [ 741.127579] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] [ 741.127579] env[62600]: INFO nova.compute.manager [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Terminating instance [ 741.127579] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Acquiring lock "refresh_cache-68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.362453] env[62600]: DEBUG nova.network.neutron [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 741.461425] env[62600]: DEBUG nova.network.neutron [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.485338] env[62600]: DEBUG nova.network.neutron [req-c71a984d-569f-4806-a85d-d9fe748594f3 req-da341b1c-dd18-4f57-877b-f811e811b181 service nova] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 741.567970] env[62600]: DEBUG nova.network.neutron [req-c71a984d-569f-4806-a85d-d9fe748594f3 req-da341b1c-dd18-4f57-877b-f811e811b181 service nova] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.705874] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d94ebd-8318-4ae9-a965-3545be986108 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.714037] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567959e2-c929-4489-9709-ad4646d9151b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.742708] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-591cf124-ddc1-4266-85ba-3fb8642bbcf0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.749864] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5224cad-3e21-4daa-81ce-f731599b9eb9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.762770] env[62600]: DEBUG nova.compute.provider_tree [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.964121] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Releasing lock "refresh_cache-c93cb228-0adb-423c-854f-73354cdb7012" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.964368] env[62600]: DEBUG nova.compute.manager [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 741.964551] env[62600]: DEBUG nova.compute.manager [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 741.964716] env[62600]: DEBUG nova.network.neutron [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 741.979918] env[62600]: DEBUG nova.network.neutron [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 742.072922] env[62600]: DEBUG oslo_concurrency.lockutils [req-c71a984d-569f-4806-a85d-d9fe748594f3 req-da341b1c-dd18-4f57-877b-f811e811b181 service nova] Releasing lock "refresh_cache-68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.073050] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Acquired lock "refresh_cache-68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.073163] env[62600]: DEBUG nova.network.neutron [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 742.265974] env[62600]: DEBUG nova.scheduler.client.report [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 742.482517] env[62600]: DEBUG nova.network.neutron [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.606040] env[62600]: DEBUG nova.network.neutron [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 742.771543] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.936s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.772238] env[62600]: ERROR nova.compute.manager [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7714dcfd-c51e-4fd3-ae24-8fed6b64d011, please check neutron logs for more information. [ 742.772238] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Traceback (most recent call last): [ 742.772238] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 742.772238] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] self.driver.spawn(context, instance, image_meta, [ 742.772238] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 742.772238] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] self._vmops.spawn(context, instance, image_meta, injected_files, [ 742.772238] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 742.772238] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] vm_ref = self.build_virtual_machine(instance, [ 742.772238] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 742.772238] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] vif_infos = vmwarevif.get_vif_info(self._session, [ 742.772238] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 742.772635] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] for vif in network_info: [ 742.772635] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 742.772635] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] return self._sync_wrapper(fn, *args, **kwargs) [ 742.772635] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 742.772635] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] self.wait() [ 742.772635] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 742.772635] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] self[:] = self._gt.wait() [ 742.772635] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 742.772635] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] return self._exit_event.wait() [ 742.772635] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 742.772635] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] result = hub.switch() [ 742.772635] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 742.772635] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] return self.greenlet.switch() [ 742.772965] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.772965] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] result = function(*args, **kwargs) [ 742.772965] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 742.772965] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] return func(*args, **kwargs) [ 742.772965] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 742.772965] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] raise e [ 742.772965] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 742.772965] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] nwinfo = self.network_api.allocate_for_instance( [ 742.772965] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 742.772965] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] created_port_ids = self._update_ports_for_instance( [ 742.772965] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 742.772965] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] with excutils.save_and_reraise_exception(): [ 742.772965] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.773309] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] self.force_reraise() [ 742.773309] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.773309] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] raise self.value [ 742.773309] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 742.773309] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] updated_port = self._update_port( [ 742.773309] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.773309] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] _ensure_no_port_binding_failure(port) [ 742.773309] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.773309] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] raise exception.PortBindingFailed(port_id=port['id']) [ 742.773309] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] nova.exception.PortBindingFailed: Binding failed for port 7714dcfd-c51e-4fd3-ae24-8fed6b64d011, please check neutron logs for more information. [ 742.773309] env[62600]: ERROR nova.compute.manager [instance: e3416076-1d11-47e4-9f76-9a64d7259166] [ 742.773918] env[62600]: DEBUG nova.compute.utils [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Binding failed for port 7714dcfd-c51e-4fd3-ae24-8fed6b64d011, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 742.778251] env[62600]: DEBUG nova.compute.manager [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Build of instance e3416076-1d11-47e4-9f76-9a64d7259166 was re-scheduled: Binding failed for port 7714dcfd-c51e-4fd3-ae24-8fed6b64d011, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 742.778710] env[62600]: DEBUG nova.compute.manager [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 742.778947] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Acquiring lock "refresh_cache-e3416076-1d11-47e4-9f76-9a64d7259166" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.779110] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Acquired lock "refresh_cache-e3416076-1d11-47e4-9f76-9a64d7259166" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.779271] env[62600]: DEBUG nova.network.neutron [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 742.780731] env[62600]: DEBUG oslo_concurrency.lockutils [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.268s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.785827] env[62600]: INFO nova.compute.claims [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 742.810224] env[62600]: DEBUG nova.network.neutron [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.972764] env[62600]: DEBUG nova.compute.manager [req-0acd40ca-b7ae-4590-93a3-dc0802a313a3 req-625626ef-4f32-4fce-86bc-f8e84a245011 service nova] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Received event network-vif-deleted-2f5a8c89-4486-4877-bb39-c7a87408ecc4 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 742.984937] env[62600]: INFO nova.compute.manager [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: c93cb228-0adb-423c-854f-73354cdb7012] Took 1.02 seconds to deallocate network for instance. [ 743.297140] env[62600]: DEBUG nova.network.neutron [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 743.312300] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Releasing lock "refresh_cache-68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.312546] env[62600]: DEBUG nova.compute.manager [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 743.312731] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 743.313010] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d9454a4c-cd1d-470c-a5a9-fb6c4c2b86f0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.321662] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece79e57-36d9-4619-ab65-b11ce1a92ed6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.344046] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9 could not be found. [ 743.344253] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 743.344434] env[62600]: INFO nova.compute.manager [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 743.344675] env[62600]: DEBUG oslo.service.loopingcall [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 743.344886] env[62600]: DEBUG nova.compute.manager [-] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 743.344979] env[62600]: DEBUG nova.network.neutron [-] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 743.384547] env[62600]: DEBUG nova.network.neutron [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.414544] env[62600]: DEBUG nova.network.neutron [-] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 743.887386] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Releasing lock "refresh_cache-e3416076-1d11-47e4-9f76-9a64d7259166" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.887661] env[62600]: DEBUG nova.compute.manager [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 743.887829] env[62600]: DEBUG nova.compute.manager [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 743.887988] env[62600]: DEBUG nova.network.neutron [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 743.905039] env[62600]: DEBUG nova.network.neutron [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 744.017022] env[62600]: INFO nova.scheduler.client.report [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Deleted allocations for instance c93cb228-0adb-423c-854f-73354cdb7012 [ 744.080137] env[62600]: DEBUG nova.network.neutron [-] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.114857] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2075e1a8-c6b8-41e9-98e7-0b564b139517 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.123036] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b794a3-a17b-436a-bc3c-06c4bb98eb74 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.154173] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b799fa7-cc9e-4360-80f3-42ac72dc8c64 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.161776] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f1f7282-1188-4775-8c1c-65fc4f56072f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.174545] env[62600]: DEBUG nova.compute.provider_tree [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.407746] env[62600]: DEBUG nova.network.neutron [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.525310] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84ee320b-9f23-4cf8-a52a-cf8ae29c44c6 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "c93cb228-0adb-423c-854f-73354cdb7012" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.671s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.581865] env[62600]: INFO nova.compute.manager [-] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Took 1.24 seconds to deallocate network for instance. [ 744.584065] env[62600]: DEBUG nova.compute.claims [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 744.584252] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.678110] env[62600]: DEBUG nova.scheduler.client.report [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 744.910296] env[62600]: INFO nova.compute.manager [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] [instance: e3416076-1d11-47e4-9f76-9a64d7259166] Took 1.02 seconds to deallocate network for instance. [ 745.028713] env[62600]: DEBUG nova.compute.manager [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 745.184048] env[62600]: DEBUG oslo_concurrency.lockutils [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.403s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.184177] env[62600]: DEBUG nova.compute.manager [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 745.186603] env[62600]: DEBUG oslo_concurrency.lockutils [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.671s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.550287] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.691568] env[62600]: DEBUG nova.compute.utils [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 745.693042] env[62600]: DEBUG nova.compute.manager [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 745.693227] env[62600]: DEBUG nova.network.neutron [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 745.737278] env[62600]: DEBUG nova.policy [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e7f1ac1631bf4033ae465a2a5e9aeef4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3c55b89b819d4fae9114fb120dccbed7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 745.961727] env[62600]: INFO nova.scheduler.client.report [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Deleted allocations for instance e3416076-1d11-47e4-9f76-9a64d7259166 [ 746.038195] env[62600]: DEBUG nova.network.neutron [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Successfully created port: e41ea304-1712-47b9-b8e8-e4faa3600e83 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 746.081799] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b2acbd5-25b9-42fc-b775-e9d796c58a96 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.091178] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-437b3b74-ccc8-4a62-89d8-f19d50fca4cb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.120906] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a18db4b0-6b69-4d13-8c06-83b14e011aaf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.128519] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-073cedf3-0aca-4ec9-8e06-53cfb5f3762c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.141401] env[62600]: DEBUG nova.compute.provider_tree [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.199296] env[62600]: DEBUG nova.compute.manager [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 746.423324] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "af7036df-b9f0-4ce6-962a-1edd7c1ea211" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.423324] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "af7036df-b9f0-4ce6-962a-1edd7c1ea211" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.475020] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c2c0d5b-61df-4160-bd63-0983dfd44fbf tempest-ServerActionsTestJSON-827359471 tempest-ServerActionsTestJSON-827359471-project-member] Lock "e3416076-1d11-47e4-9f76-9a64d7259166" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.644388] env[62600]: DEBUG nova.scheduler.client.report [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 746.712842] env[62600]: DEBUG nova.compute.manager [req-cc11bb2e-5021-4ce6-8e71-bcbd0a6c37be req-fb8f295f-9cc0-4ac0-8095-1ee01191c3ba service nova] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Received event network-changed-e41ea304-1712-47b9-b8e8-e4faa3600e83 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 746.713115] env[62600]: DEBUG nova.compute.manager [req-cc11bb2e-5021-4ce6-8e71-bcbd0a6c37be req-fb8f295f-9cc0-4ac0-8095-1ee01191c3ba service nova] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Refreshing instance network info cache due to event network-changed-e41ea304-1712-47b9-b8e8-e4faa3600e83. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 746.713321] env[62600]: DEBUG oslo_concurrency.lockutils [req-cc11bb2e-5021-4ce6-8e71-bcbd0a6c37be req-fb8f295f-9cc0-4ac0-8095-1ee01191c3ba service nova] Acquiring lock "refresh_cache-606becc8-ff43-469c-aca6-c35d67614960" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.713511] env[62600]: DEBUG oslo_concurrency.lockutils [req-cc11bb2e-5021-4ce6-8e71-bcbd0a6c37be req-fb8f295f-9cc0-4ac0-8095-1ee01191c3ba service nova] Acquired lock "refresh_cache-606becc8-ff43-469c-aca6-c35d67614960" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.713667] env[62600]: DEBUG nova.network.neutron [req-cc11bb2e-5021-4ce6-8e71-bcbd0a6c37be req-fb8f295f-9cc0-4ac0-8095-1ee01191c3ba service nova] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Refreshing network info cache for port e41ea304-1712-47b9-b8e8-e4faa3600e83 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 746.894173] env[62600]: ERROR nova.compute.manager [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e41ea304-1712-47b9-b8e8-e4faa3600e83, please check neutron logs for more information. [ 746.894173] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 746.894173] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.894173] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 746.894173] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 746.894173] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 746.894173] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 746.894173] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 746.894173] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.894173] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 746.894173] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.894173] env[62600]: ERROR nova.compute.manager raise self.value [ 746.894173] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 746.894173] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 746.894173] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.894173] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 746.894744] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.894744] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 746.894744] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e41ea304-1712-47b9-b8e8-e4faa3600e83, please check neutron logs for more information. [ 746.894744] env[62600]: ERROR nova.compute.manager [ 746.894744] env[62600]: Traceback (most recent call last): [ 746.894744] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 746.894744] env[62600]: listener.cb(fileno) [ 746.894744] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.894744] env[62600]: result = function(*args, **kwargs) [ 746.894744] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 746.894744] env[62600]: return func(*args, **kwargs) [ 746.894744] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.894744] env[62600]: raise e [ 746.894744] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.894744] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 746.894744] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 746.894744] env[62600]: created_port_ids = self._update_ports_for_instance( [ 746.894744] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 746.894744] env[62600]: with excutils.save_and_reraise_exception(): [ 746.894744] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.894744] env[62600]: self.force_reraise() [ 746.894744] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.894744] env[62600]: raise self.value [ 746.894744] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 746.894744] env[62600]: updated_port = self._update_port( [ 746.894744] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.894744] env[62600]: _ensure_no_port_binding_failure(port) [ 746.894744] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.894744] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 746.895698] env[62600]: nova.exception.PortBindingFailed: Binding failed for port e41ea304-1712-47b9-b8e8-e4faa3600e83, please check neutron logs for more information. [ 746.895698] env[62600]: Removing descriptor: 16 [ 746.978222] env[62600]: DEBUG nova.compute.manager [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 747.150020] env[62600]: DEBUG oslo_concurrency.lockutils [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.963s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.150625] env[62600]: ERROR nova.compute.manager [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c24a86a9-b949-4ec9-b539-4c2f1158a89d, please check neutron logs for more information. [ 747.150625] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Traceback (most recent call last): [ 747.150625] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 747.150625] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] self.driver.spawn(context, instance, image_meta, [ 747.150625] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 747.150625] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] self._vmops.spawn(context, instance, image_meta, injected_files, [ 747.150625] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 747.150625] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] vm_ref = self.build_virtual_machine(instance, [ 747.150625] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 747.150625] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] vif_infos = vmwarevif.get_vif_info(self._session, [ 747.150625] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 747.151177] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] for vif in network_info: [ 747.151177] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 747.151177] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] return self._sync_wrapper(fn, *args, **kwargs) [ 747.151177] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 747.151177] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] self.wait() [ 747.151177] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 747.151177] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] self[:] = self._gt.wait() [ 747.151177] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 747.151177] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] return self._exit_event.wait() [ 747.151177] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 747.151177] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] result = hub.switch() [ 747.151177] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 747.151177] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] return self.greenlet.switch() [ 747.151812] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.151812] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] result = function(*args, **kwargs) [ 747.151812] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 747.151812] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] return func(*args, **kwargs) [ 747.151812] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 747.151812] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] raise e [ 747.151812] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.151812] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] nwinfo = self.network_api.allocate_for_instance( [ 747.151812] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 747.151812] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] created_port_ids = self._update_ports_for_instance( [ 747.151812] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 747.151812] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] with excutils.save_and_reraise_exception(): [ 747.151812] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.152386] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] self.force_reraise() [ 747.152386] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.152386] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] raise self.value [ 747.152386] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 747.152386] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] updated_port = self._update_port( [ 747.152386] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.152386] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] _ensure_no_port_binding_failure(port) [ 747.152386] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.152386] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] raise exception.PortBindingFailed(port_id=port['id']) [ 747.152386] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] nova.exception.PortBindingFailed: Binding failed for port c24a86a9-b949-4ec9-b539-4c2f1158a89d, please check neutron logs for more information. [ 747.152386] env[62600]: ERROR nova.compute.manager [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] [ 747.152869] env[62600]: DEBUG nova.compute.utils [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Binding failed for port c24a86a9-b949-4ec9-b539-4c2f1158a89d, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 747.152869] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.083s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.155894] env[62600]: DEBUG nova.compute.manager [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Build of instance 877f5b67-68f8-48b0-8e12-622d9b50a390 was re-scheduled: Binding failed for port c24a86a9-b949-4ec9-b539-4c2f1158a89d, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 747.156143] env[62600]: DEBUG nova.compute.manager [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 747.156303] env[62600]: DEBUG oslo_concurrency.lockutils [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Acquiring lock "refresh_cache-877f5b67-68f8-48b0-8e12-622d9b50a390" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.156448] env[62600]: DEBUG oslo_concurrency.lockutils [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Acquired lock "refresh_cache-877f5b67-68f8-48b0-8e12-622d9b50a390" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.156607] env[62600]: DEBUG nova.network.neutron [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 747.208186] env[62600]: DEBUG nova.compute.manager [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 747.232650] env[62600]: DEBUG nova.network.neutron [req-cc11bb2e-5021-4ce6-8e71-bcbd0a6c37be req-fb8f295f-9cc0-4ac0-8095-1ee01191c3ba service nova] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 747.241348] env[62600]: DEBUG nova.virt.hardware [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 747.241583] env[62600]: DEBUG nova.virt.hardware [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 747.241738] env[62600]: DEBUG nova.virt.hardware [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 747.241915] env[62600]: DEBUG nova.virt.hardware [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 747.242077] env[62600]: DEBUG nova.virt.hardware [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 747.242227] env[62600]: DEBUG nova.virt.hardware [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 747.242432] env[62600]: DEBUG nova.virt.hardware [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 747.242589] env[62600]: DEBUG nova.virt.hardware [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 747.242749] env[62600]: DEBUG nova.virt.hardware [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 747.242902] env[62600]: DEBUG nova.virt.hardware [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 747.243837] env[62600]: DEBUG nova.virt.hardware [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 747.247117] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f00db90-317b-45b1-b2e2-7155ca3a8e07 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.254945] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa62c1a-2ff5-4528-a53b-67279afa4e58 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.271291] env[62600]: ERROR nova.compute.manager [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e41ea304-1712-47b9-b8e8-e4faa3600e83, please check neutron logs for more information. [ 747.271291] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] Traceback (most recent call last): [ 747.271291] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 747.271291] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] yield resources [ 747.271291] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 747.271291] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] self.driver.spawn(context, instance, image_meta, [ 747.271291] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 747.271291] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] self._vmops.spawn(context, instance, image_meta, injected_files, [ 747.271291] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 747.271291] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] vm_ref = self.build_virtual_machine(instance, [ 747.271291] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 747.271864] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] vif_infos = vmwarevif.get_vif_info(self._session, [ 747.271864] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 747.271864] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] for vif in network_info: [ 747.271864] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 747.271864] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] return self._sync_wrapper(fn, *args, **kwargs) [ 747.271864] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 747.271864] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] self.wait() [ 747.271864] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 747.271864] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] self[:] = self._gt.wait() [ 747.271864] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 747.271864] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] return self._exit_event.wait() [ 747.271864] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 747.271864] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] current.throw(*self._exc) [ 747.272368] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.272368] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] result = function(*args, **kwargs) [ 747.272368] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 747.272368] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] return func(*args, **kwargs) [ 747.272368] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 747.272368] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] raise e [ 747.272368] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.272368] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] nwinfo = self.network_api.allocate_for_instance( [ 747.272368] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 747.272368] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] created_port_ids = self._update_ports_for_instance( [ 747.272368] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 747.272368] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] with excutils.save_and_reraise_exception(): [ 747.272368] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.272916] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] self.force_reraise() [ 747.272916] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.272916] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] raise self.value [ 747.272916] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 747.272916] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] updated_port = self._update_port( [ 747.272916] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.272916] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] _ensure_no_port_binding_failure(port) [ 747.272916] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.272916] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] raise exception.PortBindingFailed(port_id=port['id']) [ 747.272916] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] nova.exception.PortBindingFailed: Binding failed for port e41ea304-1712-47b9-b8e8-e4faa3600e83, please check neutron logs for more information. [ 747.272916] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] [ 747.272916] env[62600]: INFO nova.compute.manager [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Terminating instance [ 747.273766] env[62600]: DEBUG oslo_concurrency.lockutils [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Acquiring lock "refresh_cache-606becc8-ff43-469c-aca6-c35d67614960" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.323048] env[62600]: DEBUG nova.network.neutron [req-cc11bb2e-5021-4ce6-8e71-bcbd0a6c37be req-fb8f295f-9cc0-4ac0-8095-1ee01191c3ba service nova] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.505456] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.688461] env[62600]: DEBUG nova.network.neutron [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 747.810156] env[62600]: DEBUG nova.network.neutron [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.827049] env[62600]: DEBUG oslo_concurrency.lockutils [req-cc11bb2e-5021-4ce6-8e71-bcbd0a6c37be req-fb8f295f-9cc0-4ac0-8095-1ee01191c3ba service nova] Releasing lock "refresh_cache-606becc8-ff43-469c-aca6-c35d67614960" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.827873] env[62600]: DEBUG oslo_concurrency.lockutils [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Acquired lock "refresh_cache-606becc8-ff43-469c-aca6-c35d67614960" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.828088] env[62600]: DEBUG nova.network.neutron [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 748.026716] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86eba107-27a9-445c-806a-169b7d4ffd5e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.035767] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b84740-38f7-4683-85f3-2e5ba93fcc0e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.064866] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95502b59-ffde-4ff1-b28c-07bc9fc26f1d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.072612] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b170430-3f3c-4cee-a4ea-856a1fafb89b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.085744] env[62600]: DEBUG nova.compute.provider_tree [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.316825] env[62600]: DEBUG oslo_concurrency.lockutils [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Releasing lock "refresh_cache-877f5b67-68f8-48b0-8e12-622d9b50a390" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.316825] env[62600]: DEBUG nova.compute.manager [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 748.316825] env[62600]: DEBUG nova.compute.manager [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 748.316825] env[62600]: DEBUG nova.network.neutron [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 748.332426] env[62600]: DEBUG nova.network.neutron [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 748.351839] env[62600]: DEBUG nova.network.neutron [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 748.473878] env[62600]: DEBUG nova.network.neutron [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.588651] env[62600]: DEBUG nova.scheduler.client.report [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.739244] env[62600]: DEBUG nova.compute.manager [req-5a1a4bce-08bf-4c05-b0ba-8d557d9ec302 req-39557cda-86a9-443c-81be-9f72da0101f4 service nova] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Received event network-vif-deleted-e41ea304-1712-47b9-b8e8-e4faa3600e83 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 748.836023] env[62600]: DEBUG nova.network.neutron [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.978130] env[62600]: DEBUG oslo_concurrency.lockutils [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Releasing lock "refresh_cache-606becc8-ff43-469c-aca6-c35d67614960" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.978130] env[62600]: DEBUG nova.compute.manager [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 748.978130] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 748.978130] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5be2975f-ef62-4d52-8015-b634779ead07 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.986657] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f831af-8365-491a-8412-437f7e94f6eb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.009253] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 606becc8-ff43-469c-aca6-c35d67614960 could not be found. [ 749.009485] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 749.009667] env[62600]: INFO nova.compute.manager [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Took 0.03 seconds to destroy the instance on the hypervisor. [ 749.009898] env[62600]: DEBUG oslo.service.loopingcall [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 749.010167] env[62600]: DEBUG nova.compute.manager [-] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 749.010294] env[62600]: DEBUG nova.network.neutron [-] [instance: 606becc8-ff43-469c-aca6-c35d67614960] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 749.025618] env[62600]: DEBUG nova.network.neutron [-] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 749.093889] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.941s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.094591] env[62600]: ERROR nova.compute.manager [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9d61a4cd-49a3-4f42-8a41-7d5aecfa0db3, please check neutron logs for more information. [ 749.094591] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Traceback (most recent call last): [ 749.094591] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 749.094591] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] self.driver.spawn(context, instance, image_meta, [ 749.094591] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 749.094591] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 749.094591] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 749.094591] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] vm_ref = self.build_virtual_machine(instance, [ 749.094591] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 749.094591] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] vif_infos = vmwarevif.get_vif_info(self._session, [ 749.094591] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 749.094973] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] for vif in network_info: [ 749.094973] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 749.094973] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] return self._sync_wrapper(fn, *args, **kwargs) [ 749.094973] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 749.094973] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] self.wait() [ 749.094973] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 749.094973] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] self[:] = self._gt.wait() [ 749.094973] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 749.094973] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] return self._exit_event.wait() [ 749.094973] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 749.094973] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] current.throw(*self._exc) [ 749.094973] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 749.094973] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] result = function(*args, **kwargs) [ 749.095421] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 749.095421] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] return func(*args, **kwargs) [ 749.095421] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 749.095421] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] raise e [ 749.095421] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 749.095421] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] nwinfo = self.network_api.allocate_for_instance( [ 749.095421] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 749.095421] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] created_port_ids = self._update_ports_for_instance( [ 749.095421] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 749.095421] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] with excutils.save_and_reraise_exception(): [ 749.095421] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 749.095421] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] self.force_reraise() [ 749.095421] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 749.095803] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] raise self.value [ 749.095803] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 749.095803] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] updated_port = self._update_port( [ 749.095803] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 749.095803] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] _ensure_no_port_binding_failure(port) [ 749.095803] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 749.095803] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] raise exception.PortBindingFailed(port_id=port['id']) [ 749.095803] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] nova.exception.PortBindingFailed: Binding failed for port 9d61a4cd-49a3-4f42-8a41-7d5aecfa0db3, please check neutron logs for more information. [ 749.095803] env[62600]: ERROR nova.compute.manager [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] [ 749.095803] env[62600]: DEBUG nova.compute.utils [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Binding failed for port 9d61a4cd-49a3-4f42-8a41-7d5aecfa0db3, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 749.096615] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.551s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.098674] env[62600]: INFO nova.compute.claims [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 749.101917] env[62600]: DEBUG nova.compute.manager [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Build of instance d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d was re-scheduled: Binding failed for port 9d61a4cd-49a3-4f42-8a41-7d5aecfa0db3, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 749.101917] env[62600]: DEBUG nova.compute.manager [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 749.102781] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "refresh_cache-d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.102781] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquired lock "refresh_cache-d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.102781] env[62600]: DEBUG nova.network.neutron [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 749.338794] env[62600]: INFO nova.compute.manager [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] [instance: 877f5b67-68f8-48b0-8e12-622d9b50a390] Took 1.02 seconds to deallocate network for instance. [ 749.529520] env[62600]: DEBUG nova.network.neutron [-] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.624818] env[62600]: DEBUG nova.network.neutron [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 749.709301] env[62600]: DEBUG nova.network.neutron [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.891626] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Acquiring lock "c2c20e5f-6c0e-4989-bc36-ff4a27d5c534" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.891626] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Lock "c2c20e5f-6c0e-4989-bc36-ff4a27d5c534" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.030965] env[62600]: INFO nova.compute.manager [-] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Took 1.02 seconds to deallocate network for instance. [ 750.033839] env[62600]: DEBUG nova.compute.claims [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 750.034035] env[62600]: DEBUG oslo_concurrency.lockutils [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.214483] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Releasing lock "refresh_cache-d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.214483] env[62600]: DEBUG nova.compute.manager [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 750.214483] env[62600]: DEBUG nova.compute.manager [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 750.214483] env[62600]: DEBUG nova.network.neutron [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 750.229557] env[62600]: DEBUG nova.network.neutron [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 750.366811] env[62600]: INFO nova.scheduler.client.report [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Deleted allocations for instance 877f5b67-68f8-48b0-8e12-622d9b50a390 [ 750.429516] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0c980ea-6a8c-48f0-b822-5284851e3232 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.437464] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2620b81-a421-4a73-be77-63eb46861e52 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.468419] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9beb37d-5461-40c1-8ca6-919682d0e307 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.475631] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdee8cc6-d5f0-42bc-b2b2-72300f9584bc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.488583] env[62600]: DEBUG nova.compute.provider_tree [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.732301] env[62600]: DEBUG nova.network.neutron [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.875920] env[62600]: DEBUG oslo_concurrency.lockutils [None req-35429d74-53c2-46b7-ba5f-311375ce10ca tempest-ServerActionsTestOtherB-267238380 tempest-ServerActionsTestOtherB-267238380-project-member] Lock "877f5b67-68f8-48b0-8e12-622d9b50a390" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.427s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.991422] env[62600]: DEBUG nova.scheduler.client.report [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 751.235794] env[62600]: INFO nova.compute.manager [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d] Took 1.02 seconds to deallocate network for instance. [ 751.378658] env[62600]: DEBUG nova.compute.manager [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 751.496555] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.400s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.496948] env[62600]: DEBUG nova.compute.manager [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 751.499634] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.414s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.906285] env[62600]: DEBUG oslo_concurrency.lockutils [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.010300] env[62600]: DEBUG nova.compute.utils [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 752.015537] env[62600]: DEBUG nova.compute.manager [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 752.015728] env[62600]: DEBUG nova.network.neutron [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 752.085415] env[62600]: DEBUG nova.policy [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '646761ca4bfb4dd7bc577ad4f1f1794c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '529e31e208264cc7981374c676b4e7ab', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 752.306837] env[62600]: INFO nova.scheduler.client.report [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Deleted allocations for instance d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d [ 752.472142] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-133b831c-05c0-41e8-a3ad-ff4595a2f805 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.483517] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ad989c-4605-4158-ab5d-4044d0612975 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.520192] env[62600]: DEBUG nova.compute.manager [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 752.524149] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b31d98-7918-4f72-9da9-4455c0907753 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.538294] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8364732-5294-4249-b96c-81f0ce5159a8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.550353] env[62600]: DEBUG nova.compute.provider_tree [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.610254] env[62600]: DEBUG nova.network.neutron [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Successfully created port: 42193f90-0f3e-4b54-8eb0-f5ec95237a03 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 752.809390] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4b19041c-9e43-40f0-9f40-e350ce0efc55 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "d94d2c98-bbbb-4bdf-a4d4-603aa7635a7d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.710s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.029919] env[62600]: INFO nova.virt.block_device [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Booting with volume a4d4a217-a37c-4a66-8b5e-80db1d36d524 at /dev/sda [ 753.059289] env[62600]: DEBUG nova.scheduler.client.report [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 753.119557] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c7854f38-0f38-46e1-8e12-b1616d751864 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.132519] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec5a241-d4ca-41e9-a8fe-72f35f7a370e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.162774] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8545acfb-6e3b-4b21-93e9-0adac94a9350 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.171912] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab25dd4-541d-4584-ac6f-21f8cb38b96e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.195334] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b811c157-5da4-4cf9-a2e3-00b7dee128cd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.200027] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d3a937-b69b-4da4-bf2f-c55c03904353 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.213739] env[62600]: DEBUG nova.virt.block_device [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Updating existing volume attachment record: 977db38e-80cf-4b67-ad1c-6d7f603e6997 {{(pid=62600) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 753.316410] env[62600]: DEBUG nova.compute.manager [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 753.450514] env[62600]: DEBUG nova.compute.manager [req-a8a08d4d-7169-4123-ad36-8f9acdcc9222 req-5121bd57-d19f-46c1-932e-ef39c53c125b service nova] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Received event network-changed-42193f90-0f3e-4b54-8eb0-f5ec95237a03 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 753.450514] env[62600]: DEBUG nova.compute.manager [req-a8a08d4d-7169-4123-ad36-8f9acdcc9222 req-5121bd57-d19f-46c1-932e-ef39c53c125b service nova] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Refreshing instance network info cache due to event network-changed-42193f90-0f3e-4b54-8eb0-f5ec95237a03. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 753.450728] env[62600]: DEBUG oslo_concurrency.lockutils [req-a8a08d4d-7169-4123-ad36-8f9acdcc9222 req-5121bd57-d19f-46c1-932e-ef39c53c125b service nova] Acquiring lock "refresh_cache-014899cb-db4c-4444-abe1-b0129c0d7db9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.450728] env[62600]: DEBUG oslo_concurrency.lockutils [req-a8a08d4d-7169-4123-ad36-8f9acdcc9222 req-5121bd57-d19f-46c1-932e-ef39c53c125b service nova] Acquired lock "refresh_cache-014899cb-db4c-4444-abe1-b0129c0d7db9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.451010] env[62600]: DEBUG nova.network.neutron [req-a8a08d4d-7169-4123-ad36-8f9acdcc9222 req-5121bd57-d19f-46c1-932e-ef39c53c125b service nova] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Refreshing network info cache for port 42193f90-0f3e-4b54-8eb0-f5ec95237a03 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 753.568842] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.069s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.569270] env[62600]: ERROR nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3388af72-15ab-44e3-ab44-8a8ad11950a9, please check neutron logs for more information. [ 753.569270] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Traceback (most recent call last): [ 753.569270] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 753.569270] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] self.driver.spawn(context, instance, image_meta, [ 753.569270] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 753.569270] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] self._vmops.spawn(context, instance, image_meta, injected_files, [ 753.569270] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 753.569270] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] vm_ref = self.build_virtual_machine(instance, [ 753.569270] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 753.569270] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] vif_infos = vmwarevif.get_vif_info(self._session, [ 753.569270] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 753.572502] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] for vif in network_info: [ 753.572502] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 753.572502] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] return self._sync_wrapper(fn, *args, **kwargs) [ 753.572502] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 753.572502] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] self.wait() [ 753.572502] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 753.572502] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] self[:] = self._gt.wait() [ 753.572502] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 753.572502] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] return self._exit_event.wait() [ 753.572502] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 753.572502] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] current.throw(*self._exc) [ 753.572502] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.572502] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] result = function(*args, **kwargs) [ 753.572863] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 753.572863] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] return func(*args, **kwargs) [ 753.572863] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.572863] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] raise e [ 753.572863] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.572863] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] nwinfo = self.network_api.allocate_for_instance( [ 753.572863] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 753.572863] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] created_port_ids = self._update_ports_for_instance( [ 753.572863] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 753.572863] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] with excutils.save_and_reraise_exception(): [ 753.572863] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.572863] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] self.force_reraise() [ 753.572863] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.573228] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] raise self.value [ 753.573228] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 753.573228] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] updated_port = self._update_port( [ 753.573228] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.573228] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] _ensure_no_port_binding_failure(port) [ 753.573228] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.573228] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] raise exception.PortBindingFailed(port_id=port['id']) [ 753.573228] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] nova.exception.PortBindingFailed: Binding failed for port 3388af72-15ab-44e3-ab44-8a8ad11950a9, please check neutron logs for more information. [ 753.573228] env[62600]: ERROR nova.compute.manager [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] [ 753.573228] env[62600]: DEBUG nova.compute.utils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Binding failed for port 3388af72-15ab-44e3-ab44-8a8ad11950a9, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 753.573502] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.663s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.575104] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Build of instance f989a4e4-c3c6-4cb5-9464-cbfb9d66c202 was re-scheduled: Binding failed for port 3388af72-15ab-44e3-ab44-8a8ad11950a9, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 753.575633] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 753.575951] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquiring lock "refresh_cache-f989a4e4-c3c6-4cb5-9464-cbfb9d66c202" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.576190] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquired lock "refresh_cache-f989a4e4-c3c6-4cb5-9464-cbfb9d66c202" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.576416] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 753.637334] env[62600]: ERROR nova.compute.manager [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 42193f90-0f3e-4b54-8eb0-f5ec95237a03, please check neutron logs for more information. [ 753.637334] env[62600]: ERROR nova.compute.manager Traceback (most recent call last): [ 753.637334] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.637334] env[62600]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 753.637334] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 753.637334] env[62600]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 753.637334] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 753.637334] env[62600]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 753.637334] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.637334] env[62600]: ERROR nova.compute.manager self.force_reraise() [ 753.637334] env[62600]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.637334] env[62600]: ERROR nova.compute.manager raise self.value [ 753.637334] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 753.637334] env[62600]: ERROR nova.compute.manager updated_port = self._update_port( [ 753.637334] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.637334] env[62600]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 753.637967] env[62600]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.637967] env[62600]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 753.637967] env[62600]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 42193f90-0f3e-4b54-8eb0-f5ec95237a03, please check neutron logs for more information. [ 753.637967] env[62600]: ERROR nova.compute.manager [ 753.637967] env[62600]: Traceback (most recent call last): [ 753.637967] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 753.637967] env[62600]: listener.cb(fileno) [ 753.637967] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.637967] env[62600]: result = function(*args, **kwargs) [ 753.637967] env[62600]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 753.637967] env[62600]: return func(*args, **kwargs) [ 753.637967] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 753.637967] env[62600]: raise e [ 753.637967] env[62600]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 753.637967] env[62600]: nwinfo = self.network_api.allocate_for_instance( [ 753.637967] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 753.637967] env[62600]: created_port_ids = self._update_ports_for_instance( [ 753.637967] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 753.637967] env[62600]: with excutils.save_and_reraise_exception(): [ 753.637967] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.637967] env[62600]: self.force_reraise() [ 753.637967] env[62600]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.637967] env[62600]: raise self.value [ 753.637967] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 753.637967] env[62600]: updated_port = self._update_port( [ 753.637967] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.637967] env[62600]: _ensure_no_port_binding_failure(port) [ 753.637967] env[62600]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.637967] env[62600]: raise exception.PortBindingFailed(port_id=port['id']) [ 753.638871] env[62600]: nova.exception.PortBindingFailed: Binding failed for port 42193f90-0f3e-4b54-8eb0-f5ec95237a03, please check neutron logs for more information. [ 753.638871] env[62600]: Removing descriptor: 15 [ 753.839959] env[62600]: DEBUG oslo_concurrency.lockutils [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.968837] env[62600]: DEBUG nova.network.neutron [req-a8a08d4d-7169-4123-ad36-8f9acdcc9222 req-5121bd57-d19f-46c1-932e-ef39c53c125b service nova] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 754.050322] env[62600]: DEBUG nova.network.neutron [req-a8a08d4d-7169-4123-ad36-8f9acdcc9222 req-5121bd57-d19f-46c1-932e-ef39c53c125b service nova] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.097241] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 754.175503] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.300359] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "9d899d96-9f4f-41d1-a368-3fde5efc110a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.300856] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "9d899d96-9f4f-41d1-a368-3fde5efc110a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.409367] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e837cb69-b530-4394-b1d8-ccd1ecbca2cf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.417119] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b99e93-1a0c-43d9-b4b1-04093bf3f5e3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.446527] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92af80a8-feea-41c7-aac5-621a7b0920d8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.453738] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1770598e-eda8-4031-9a75-6017cb8da1d8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.469309] env[62600]: DEBUG nova.compute.provider_tree [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.553364] env[62600]: DEBUG oslo_concurrency.lockutils [req-a8a08d4d-7169-4123-ad36-8f9acdcc9222 req-5121bd57-d19f-46c1-932e-ef39c53c125b service nova] Releasing lock "refresh_cache-014899cb-db4c-4444-abe1-b0129c0d7db9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.677969] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Releasing lock "refresh_cache-f989a4e4-c3c6-4cb5-9464-cbfb9d66c202" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.678346] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 754.678639] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 754.678877] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 754.695367] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 754.971989] env[62600]: DEBUG nova.scheduler.client.report [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 755.198271] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.322937] env[62600]: DEBUG nova.compute.manager [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 755.323549] env[62600]: DEBUG nova.virt.hardware [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 755.323738] env[62600]: DEBUG nova.virt.hardware [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 755.323894] env[62600]: DEBUG nova.virt.hardware [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 755.324085] env[62600]: DEBUG nova.virt.hardware [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 755.324234] env[62600]: DEBUG nova.virt.hardware [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 755.324379] env[62600]: DEBUG nova.virt.hardware [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 755.324578] env[62600]: DEBUG nova.virt.hardware [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 755.324734] env[62600]: DEBUG nova.virt.hardware [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 755.324896] env[62600]: DEBUG nova.virt.hardware [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 755.325159] env[62600]: DEBUG nova.virt.hardware [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 755.325862] env[62600]: DEBUG nova.virt.hardware [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 755.327190] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a35bff-2d0d-487e-8b6c-aac3b62bf42a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.336214] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f021406-bb9f-4b7c-9a1e-fcd796d23039 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.351922] env[62600]: ERROR nova.compute.manager [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 42193f90-0f3e-4b54-8eb0-f5ec95237a03, please check neutron logs for more information. [ 755.351922] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Traceback (most recent call last): [ 755.351922] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 755.351922] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] yield resources [ 755.351922] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 755.351922] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] self.driver.spawn(context, instance, image_meta, [ 755.351922] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 755.351922] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 755.351922] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 755.351922] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] vm_ref = self.build_virtual_machine(instance, [ 755.351922] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 755.352293] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] vif_infos = vmwarevif.get_vif_info(self._session, [ 755.352293] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 755.352293] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] for vif in network_info: [ 755.352293] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 755.352293] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] return self._sync_wrapper(fn, *args, **kwargs) [ 755.352293] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 755.352293] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] self.wait() [ 755.352293] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 755.352293] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] self[:] = self._gt.wait() [ 755.352293] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 755.352293] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] return self._exit_event.wait() [ 755.352293] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 755.352293] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] current.throw(*self._exc) [ 755.352737] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 755.352737] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] result = function(*args, **kwargs) [ 755.352737] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 755.352737] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] return func(*args, **kwargs) [ 755.352737] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 755.352737] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] raise e [ 755.352737] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 755.352737] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] nwinfo = self.network_api.allocate_for_instance( [ 755.352737] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 755.352737] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] created_port_ids = self._update_ports_for_instance( [ 755.352737] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 755.352737] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] with excutils.save_and_reraise_exception(): [ 755.352737] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.353164] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] self.force_reraise() [ 755.353164] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.353164] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] raise self.value [ 755.353164] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 755.353164] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] updated_port = self._update_port( [ 755.353164] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.353164] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] _ensure_no_port_binding_failure(port) [ 755.353164] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.353164] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] raise exception.PortBindingFailed(port_id=port['id']) [ 755.353164] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] nova.exception.PortBindingFailed: Binding failed for port 42193f90-0f3e-4b54-8eb0-f5ec95237a03, please check neutron logs for more information. [ 755.353164] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] [ 755.353164] env[62600]: INFO nova.compute.manager [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Terminating instance [ 755.354234] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Acquiring lock "refresh_cache-014899cb-db4c-4444-abe1-b0129c0d7db9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.354393] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Acquired lock "refresh_cache-014899cb-db4c-4444-abe1-b0129c0d7db9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.354562] env[62600]: DEBUG nova.network.neutron [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 755.473726] env[62600]: DEBUG nova.compute.manager [req-2e5f343e-43a7-4de2-b619-cd41a2cc183d req-73453818-49db-4fe4-b1d3-ff5cd7f89cb4 service nova] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Received event network-vif-deleted-42193f90-0f3e-4b54-8eb0-f5ec95237a03 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 755.476840] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.905s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.477460] env[62600]: ERROR nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e8122e39-0982-4178-97a6-52a3f6e5c9e8, please check neutron logs for more information. [ 755.477460] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Traceback (most recent call last): [ 755.477460] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 755.477460] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] self.driver.spawn(context, instance, image_meta, [ 755.477460] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 755.477460] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 755.477460] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 755.477460] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] vm_ref = self.build_virtual_machine(instance, [ 755.477460] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 755.477460] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] vif_infos = vmwarevif.get_vif_info(self._session, [ 755.477460] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 755.477885] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] for vif in network_info: [ 755.477885] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 755.477885] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] return self._sync_wrapper(fn, *args, **kwargs) [ 755.477885] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 755.477885] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] self.wait() [ 755.477885] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 755.477885] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] self[:] = self._gt.wait() [ 755.477885] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 755.477885] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] return self._exit_event.wait() [ 755.477885] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 755.477885] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] current.throw(*self._exc) [ 755.477885] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 755.477885] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] result = function(*args, **kwargs) [ 755.478256] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 755.478256] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] return func(*args, **kwargs) [ 755.478256] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 755.478256] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] raise e [ 755.478256] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 755.478256] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] nwinfo = self.network_api.allocate_for_instance( [ 755.478256] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 755.478256] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] created_port_ids = self._update_ports_for_instance( [ 755.478256] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 755.478256] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] with excutils.save_and_reraise_exception(): [ 755.478256] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.478256] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] self.force_reraise() [ 755.478256] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.478710] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] raise self.value [ 755.478710] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 755.478710] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] updated_port = self._update_port( [ 755.478710] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.478710] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] _ensure_no_port_binding_failure(port) [ 755.478710] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.478710] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] raise exception.PortBindingFailed(port_id=port['id']) [ 755.478710] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] nova.exception.PortBindingFailed: Binding failed for port e8122e39-0982-4178-97a6-52a3f6e5c9e8, please check neutron logs for more information. [ 755.478710] env[62600]: ERROR nova.compute.manager [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] [ 755.478710] env[62600]: DEBUG nova.compute.utils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Binding failed for port e8122e39-0982-4178-97a6-52a3f6e5c9e8, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 755.479760] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.937s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.482944] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Build of instance 9c2c8dd4-85c7-489a-95c0-669644237b8b was re-scheduled: Binding failed for port e8122e39-0982-4178-97a6-52a3f6e5c9e8, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 755.483373] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 755.483610] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquiring lock "refresh_cache-9c2c8dd4-85c7-489a-95c0-669644237b8b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.483812] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquired lock "refresh_cache-9c2c8dd4-85c7-489a-95c0-669644237b8b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.483923] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 755.701188] env[62600]: INFO nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: f989a4e4-c3c6-4cb5-9464-cbfb9d66c202] Took 1.02 seconds to deallocate network for instance. [ 755.872051] env[62600]: DEBUG nova.network.neutron [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 755.956789] env[62600]: DEBUG nova.network.neutron [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.004163] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 756.080675] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.274221] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d454a97-2bb5-4edc-ab86-1e9cbad70ab4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.281893] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e5a49a1-2e0c-4705-9838-98eae2d63bb6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.311136] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c395772a-68ab-404c-a01a-abad6b992c32 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.318487] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edf59a66-d578-4060-9ec2-c49c02808fcd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.331579] env[62600]: DEBUG nova.compute.provider_tree [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.460695] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Releasing lock "refresh_cache-014899cb-db4c-4444-abe1-b0129c0d7db9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.461336] env[62600]: DEBUG nova.compute.manager [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 756.461667] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-89dfec57-fac0-459e-9015-f266b51e6d8c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.471124] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afab1f85-e00f-46d2-8830-33572911b3fc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.493420] env[62600]: WARNING nova.virt.vmwareapi.driver [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 014899cb-db4c-4444-abe1-b0129c0d7db9 could not be found. [ 756.493787] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 756.494054] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a86859c8-e0c0-48e4-95bc-b6469e8e669c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.501267] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df2a5cc4-2454-4e57-a5bf-9dba8188b709 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.523079] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 014899cb-db4c-4444-abe1-b0129c0d7db9 could not be found. [ 756.523316] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 756.523490] env[62600]: INFO nova.compute.manager [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Took 0.06 seconds to destroy the instance on the hypervisor. [ 756.523723] env[62600]: DEBUG oslo.service.loopingcall [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 756.523939] env[62600]: DEBUG nova.compute.manager [-] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 756.524032] env[62600]: DEBUG nova.network.neutron [-] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 756.538668] env[62600]: DEBUG nova.network.neutron [-] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 756.582806] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Releasing lock "refresh_cache-9c2c8dd4-85c7-489a-95c0-669644237b8b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.583043] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 756.583223] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 756.583496] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 756.599475] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 756.732192] env[62600]: INFO nova.scheduler.client.report [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Deleted allocations for instance f989a4e4-c3c6-4cb5-9464-cbfb9d66c202 [ 756.834456] env[62600]: DEBUG nova.scheduler.client.report [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 757.041040] env[62600]: DEBUG nova.network.neutron [-] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.102417] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.239578] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Lock "f989a4e4-c3c6-4cb5-9464-cbfb9d66c202" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.027s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.339715] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.860s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.340379] env[62600]: ERROR nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 47c7b2fe-dec4-439d-b0a7-f1b14dbf1f05, please check neutron logs for more information. [ 757.340379] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] Traceback (most recent call last): [ 757.340379] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 757.340379] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] self.driver.spawn(context, instance, image_meta, [ 757.340379] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 757.340379] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] self._vmops.spawn(context, instance, image_meta, injected_files, [ 757.340379] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 757.340379] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] vm_ref = self.build_virtual_machine(instance, [ 757.340379] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 757.340379] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] vif_infos = vmwarevif.get_vif_info(self._session, [ 757.340379] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 757.340824] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] for vif in network_info: [ 757.340824] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 757.340824] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] return self._sync_wrapper(fn, *args, **kwargs) [ 757.340824] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 757.340824] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] self.wait() [ 757.340824] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 757.340824] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] self[:] = self._gt.wait() [ 757.340824] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 757.340824] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] return self._exit_event.wait() [ 757.340824] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 757.340824] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] current.throw(*self._exc) [ 757.340824] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.340824] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] result = function(*args, **kwargs) [ 757.341221] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 757.341221] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] return func(*args, **kwargs) [ 757.341221] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.341221] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] raise e [ 757.341221] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.341221] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] nwinfo = self.network_api.allocate_for_instance( [ 757.341221] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 757.341221] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] created_port_ids = self._update_ports_for_instance( [ 757.341221] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 757.341221] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] with excutils.save_and_reraise_exception(): [ 757.341221] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.341221] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] self.force_reraise() [ 757.341221] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.341622] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] raise self.value [ 757.341622] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 757.341622] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] updated_port = self._update_port( [ 757.341622] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.341622] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] _ensure_no_port_binding_failure(port) [ 757.341622] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.341622] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] raise exception.PortBindingFailed(port_id=port['id']) [ 757.341622] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] nova.exception.PortBindingFailed: Binding failed for port 47c7b2fe-dec4-439d-b0a7-f1b14dbf1f05, please check neutron logs for more information. [ 757.341622] env[62600]: ERROR nova.compute.manager [instance: 81814471-a617-4364-a140-105252ca9c04] [ 757.341622] env[62600]: DEBUG nova.compute.utils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Binding failed for port 47c7b2fe-dec4-439d-b0a7-f1b14dbf1f05, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 757.342405] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.758s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.345591] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Build of instance 81814471-a617-4364-a140-105252ca9c04 was re-scheduled: Binding failed for port 47c7b2fe-dec4-439d-b0a7-f1b14dbf1f05, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 757.347825] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 757.348108] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquiring lock "refresh_cache-81814471-a617-4364-a140-105252ca9c04" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.348298] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Acquired lock "refresh_cache-81814471-a617-4364-a140-105252ca9c04" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.348491] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 757.544487] env[62600]: INFO nova.compute.manager [-] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Took 1.02 seconds to deallocate network for instance. [ 757.608527] env[62600]: INFO nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 9c2c8dd4-85c7-489a-95c0-669644237b8b] Took 1.02 seconds to deallocate network for instance. [ 757.746034] env[62600]: DEBUG nova.compute.manager [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 757.868760] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 757.980988] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.143696] env[62600]: INFO nova.compute.manager [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Took 0.60 seconds to detach 1 volumes for instance. [ 758.146281] env[62600]: DEBUG nova.compute.claims [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Aborting claim: {{(pid=62600) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 758.146466] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.198747] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b50ea20a-63b9-42d7-be65-961c8cc18f6d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.206680] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-491aa23a-5fa4-4ce2-9afc-ae864f434b42 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.238926] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3205e73f-6a81-45fd-82cb-65c956df032b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.246261] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb8beefc-2257-4649-a86e-41793afb8553 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.263347] env[62600]: DEBUG nova.compute.provider_tree [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.275794] env[62600]: DEBUG oslo_concurrency.lockutils [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.483459] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Releasing lock "refresh_cache-81814471-a617-4364-a140-105252ca9c04" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.483768] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 758.483996] env[62600]: DEBUG nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 758.484217] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 758.502747] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 758.651386] env[62600]: INFO nova.scheduler.client.report [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Deleted allocations for instance 9c2c8dd4-85c7-489a-95c0-669644237b8b [ 758.766061] env[62600]: DEBUG nova.scheduler.client.report [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 759.005970] env[62600]: DEBUG nova.network.neutron [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.159837] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Lock "9c2c8dd4-85c7-489a-95c0-669644237b8b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.915s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.271014] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.928s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.271725] env[62600]: ERROR nova.compute.manager [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2f5a8c89-4486-4877-bb39-c7a87408ecc4, please check neutron logs for more information. [ 759.271725] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Traceback (most recent call last): [ 759.271725] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 759.271725] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] self.driver.spawn(context, instance, image_meta, [ 759.271725] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 759.271725] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 759.271725] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 759.271725] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] vm_ref = self.build_virtual_machine(instance, [ 759.271725] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 759.271725] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] vif_infos = vmwarevif.get_vif_info(self._session, [ 759.271725] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 759.272121] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] for vif in network_info: [ 759.272121] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 759.272121] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] return self._sync_wrapper(fn, *args, **kwargs) [ 759.272121] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 759.272121] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] self.wait() [ 759.272121] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 759.272121] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] self[:] = self._gt.wait() [ 759.272121] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 759.272121] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] return self._exit_event.wait() [ 759.272121] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 759.272121] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] result = hub.switch() [ 759.272121] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 759.272121] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] return self.greenlet.switch() [ 759.272537] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 759.272537] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] result = function(*args, **kwargs) [ 759.272537] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 759.272537] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] return func(*args, **kwargs) [ 759.272537] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 759.272537] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] raise e [ 759.272537] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 759.272537] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] nwinfo = self.network_api.allocate_for_instance( [ 759.272537] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 759.272537] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] created_port_ids = self._update_ports_for_instance( [ 759.272537] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 759.272537] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] with excutils.save_and_reraise_exception(): [ 759.272537] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.272925] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] self.force_reraise() [ 759.272925] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.272925] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] raise self.value [ 759.272925] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 759.272925] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] updated_port = self._update_port( [ 759.272925] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.272925] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] _ensure_no_port_binding_failure(port) [ 759.272925] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.272925] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] raise exception.PortBindingFailed(port_id=port['id']) [ 759.272925] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] nova.exception.PortBindingFailed: Binding failed for port 2f5a8c89-4486-4877-bb39-c7a87408ecc4, please check neutron logs for more information. [ 759.272925] env[62600]: ERROR nova.compute.manager [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] [ 759.273278] env[62600]: DEBUG nova.compute.utils [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Binding failed for port 2f5a8c89-4486-4877-bb39-c7a87408ecc4, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 759.273639] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.723s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.275124] env[62600]: INFO nova.compute.claims [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 759.277787] env[62600]: DEBUG nova.compute.manager [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Build of instance 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9 was re-scheduled: Binding failed for port 2f5a8c89-4486-4877-bb39-c7a87408ecc4, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 759.278215] env[62600]: DEBUG nova.compute.manager [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 759.278452] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Acquiring lock "refresh_cache-68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.278629] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Acquired lock "refresh_cache-68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.278813] env[62600]: DEBUG nova.network.neutron [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 759.508622] env[62600]: INFO nova.compute.manager [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] [instance: 81814471-a617-4364-a140-105252ca9c04] Took 1.02 seconds to deallocate network for instance. [ 759.662112] env[62600]: DEBUG nova.compute.manager [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 759.797992] env[62600]: DEBUG nova.network.neutron [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 759.884481] env[62600]: DEBUG nova.network.neutron [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.183699] env[62600]: DEBUG oslo_concurrency.lockutils [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.387106] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Releasing lock "refresh_cache-68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.387344] env[62600]: DEBUG nova.compute.manager [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 760.387516] env[62600]: DEBUG nova.compute.manager [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 760.387729] env[62600]: DEBUG nova.network.neutron [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 760.406341] env[62600]: DEBUG nova.network.neutron [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 760.533575] env[62600]: INFO nova.scheduler.client.report [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Deleted allocations for instance 81814471-a617-4364-a140-105252ca9c04 [ 760.564015] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0939fbd-6759-4a0a-9087-39ea6e746244 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.571738] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0de3d4d9-f22d-4971-8568-1dc8045aa27b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.602704] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9616a34-1495-4414-9644-675d28244ba9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.609858] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e56d439-3c7d-4f39-a848-123747dc954e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.623680] env[62600]: DEBUG nova.compute.provider_tree [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.908914] env[62600]: DEBUG nova.network.neutron [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.044367] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cbc6b89f-5300-4f1c-a4c9-79ee8a394318 tempest-ListServersNegativeTestJSON-911370175 tempest-ListServersNegativeTestJSON-911370175-project-member] Lock "81814471-a617-4364-a140-105252ca9c04" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.767s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.127042] env[62600]: DEBUG nova.scheduler.client.report [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 761.411717] env[62600]: INFO nova.compute.manager [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] [instance: 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9] Took 1.02 seconds to deallocate network for instance. [ 761.546901] env[62600]: DEBUG nova.compute.manager [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 761.632292] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.358s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.632698] env[62600]: DEBUG nova.compute.manager [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 761.635108] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.130s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.636619] env[62600]: INFO nova.compute.claims [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.069787] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.140793] env[62600]: DEBUG nova.compute.utils [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 762.143876] env[62600]: DEBUG nova.compute.manager [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 762.144052] env[62600]: DEBUG nova.network.neutron [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 762.194721] env[62600]: DEBUG nova.policy [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '76f3dcf16ac34df683548f101b258060', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2141e7e1a0884ed1b512cf12e0d01e55', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 762.433195] env[62600]: DEBUG nova.network.neutron [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Successfully created port: fe02d314-372f-43fb-8d7c-cfc33c9e74bd {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 762.450238] env[62600]: INFO nova.scheduler.client.report [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Deleted allocations for instance 68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9 [ 762.647224] env[62600]: DEBUG nova.compute.manager [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 762.964508] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ca3ed69-0316-473e-96ff-ec80ff6aec71 tempest-ServersTestMultiNic-1154643049 tempest-ServersTestMultiNic-1154643049-project-member] Lock "68a3a4a5-ee3e-4234-8577-e5c1af6fbfb9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.195s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.980277] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe2b4c91-37d7-40a6-88e1-4a92819a4602 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.989365] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dfb1222-2818-44e5-b06e-d7f4f5504f34 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.022395] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f19ea6a-f3eb-451f-a4d4-3ede71430b7f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.029691] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d36d135-722b-4372-9f18-a98a1db25fdb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.042747] env[62600]: DEBUG nova.compute.provider_tree [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.467893] env[62600]: DEBUG nova.compute.manager [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 763.548359] env[62600]: DEBUG nova.scheduler.client.report [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 763.663158] env[62600]: DEBUG nova.compute.manager [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 763.697594] env[62600]: DEBUG nova.virt.hardware [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T12:01:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='99810421-2db4-4272-82bd-17ba5175b496',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-2077855957',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 763.697871] env[62600]: DEBUG nova.virt.hardware [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 763.698047] env[62600]: DEBUG nova.virt.hardware [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 763.698234] env[62600]: DEBUG nova.virt.hardware [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 763.698379] env[62600]: DEBUG nova.virt.hardware [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 763.698523] env[62600]: DEBUG nova.virt.hardware [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 763.698726] env[62600]: DEBUG nova.virt.hardware [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 763.698941] env[62600]: DEBUG nova.virt.hardware [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 763.699337] env[62600]: DEBUG nova.virt.hardware [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 763.699539] env[62600]: DEBUG nova.virt.hardware [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 763.699700] env[62600]: DEBUG nova.virt.hardware [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 763.700560] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d72ddba-4836-47be-b380-4030f4e3b2f6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.708613] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d14faa1-e9d4-4434-b7da-307d21fa4c26 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.919993] env[62600]: DEBUG nova.compute.manager [req-e3257ae9-c46b-48fb-a286-64700fa33cd6 req-ef996d00-ed7b-4e49-b57d-468d5e3b708e service nova] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Received event network-vif-plugged-fe02d314-372f-43fb-8d7c-cfc33c9e74bd {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 763.920485] env[62600]: DEBUG oslo_concurrency.lockutils [req-e3257ae9-c46b-48fb-a286-64700fa33cd6 req-ef996d00-ed7b-4e49-b57d-468d5e3b708e service nova] Acquiring lock "de4f9304-3357-4eaa-9c94-fe28bc554086-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.920759] env[62600]: DEBUG oslo_concurrency.lockutils [req-e3257ae9-c46b-48fb-a286-64700fa33cd6 req-ef996d00-ed7b-4e49-b57d-468d5e3b708e service nova] Lock "de4f9304-3357-4eaa-9c94-fe28bc554086-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.920992] env[62600]: DEBUG oslo_concurrency.lockutils [req-e3257ae9-c46b-48fb-a286-64700fa33cd6 req-ef996d00-ed7b-4e49-b57d-468d5e3b708e service nova] Lock "de4f9304-3357-4eaa-9c94-fe28bc554086-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.921396] env[62600]: DEBUG nova.compute.manager [req-e3257ae9-c46b-48fb-a286-64700fa33cd6 req-ef996d00-ed7b-4e49-b57d-468d5e3b708e service nova] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] No waiting events found dispatching network-vif-plugged-fe02d314-372f-43fb-8d7c-cfc33c9e74bd {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 763.921441] env[62600]: WARNING nova.compute.manager [req-e3257ae9-c46b-48fb-a286-64700fa33cd6 req-ef996d00-ed7b-4e49-b57d-468d5e3b708e service nova] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Received unexpected event network-vif-plugged-fe02d314-372f-43fb-8d7c-cfc33c9e74bd for instance with vm_state building and task_state spawning. [ 764.002249] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.053180] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.418s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.053697] env[62600]: DEBUG nova.compute.manager [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 764.056607] env[62600]: DEBUG oslo_concurrency.lockutils [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.023s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.078395] env[62600]: DEBUG nova.network.neutron [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Successfully updated port: fe02d314-372f-43fb-8d7c-cfc33c9e74bd {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 764.563578] env[62600]: DEBUG nova.compute.utils [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 764.568389] env[62600]: DEBUG nova.compute.manager [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 764.568709] env[62600]: DEBUG nova.network.neutron [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 764.580766] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "refresh_cache-de4f9304-3357-4eaa-9c94-fe28bc554086" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.580766] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquired lock "refresh_cache-de4f9304-3357-4eaa-9c94-fe28bc554086" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.580766] env[62600]: DEBUG nova.network.neutron [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 764.624179] env[62600]: DEBUG nova.policy [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd3879a63b63545368a88da4574e8a4c3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '49c5a7585f2a4d7888f3240992a68955', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 764.920427] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73e643b-7517-49c1-8fc8-077c48e89fdc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.929723] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe0c1ee-3eec-4007-9063-e8a6c4571f77 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.967926] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c274ef2a-d79f-4f54-9610-9fbced63194c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.977566] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aac8e3d7-0a98-47c6-9e38-b3ba0b448a90 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.991858] env[62600]: DEBUG nova.compute.provider_tree [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.993694] env[62600]: DEBUG nova.network.neutron [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Successfully created port: 779d4e97-daa9-44ec-9ee0-91658a73bc02 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 765.069444] env[62600]: DEBUG nova.compute.manager [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 765.120267] env[62600]: DEBUG nova.network.neutron [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 765.375990] env[62600]: DEBUG nova.network.neutron [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Updating instance_info_cache with network_info: [{"id": "fe02d314-372f-43fb-8d7c-cfc33c9e74bd", "address": "fa:16:3e:86:22:ec", "network": {"id": "d946b94e-9827-4c22-af05-072fcaa3d0e6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.235", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ec39f2547f624ed0b240d084b949694a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe02d314-37", "ovs_interfaceid": "fe02d314-372f-43fb-8d7c-cfc33c9e74bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.497029] env[62600]: DEBUG nova.scheduler.client.report [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 765.579443] env[62600]: INFO nova.virt.block_device [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Booting with volume 88d3934f-d0ae-4578-af24-9abb842074b3 at /dev/sda [ 765.627257] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f6d9a470-2a67-445f-96f9-65e1b9b653b4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.637950] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e3ba050-1137-412a-923a-214565032ddb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.665313] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6bbf2c52-1721-4518-b848-576008dd3449 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.673673] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e7861c-a575-4caa-8c89-9536def5db05 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.693114] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c7c31c2-07e9-4227-911c-5f4b670ed2ed {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.699464] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2afb92aa-6c5f-436d-adf6-8fbbd611ad52 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.712318] env[62600]: DEBUG nova.virt.block_device [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Updating existing volume attachment record: f8573456-d219-47bb-87ae-26f22ee5429f {{(pid=62600) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 765.879306] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Releasing lock "refresh_cache-de4f9304-3357-4eaa-9c94-fe28bc554086" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.879727] env[62600]: DEBUG nova.compute.manager [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Instance network_info: |[{"id": "fe02d314-372f-43fb-8d7c-cfc33c9e74bd", "address": "fa:16:3e:86:22:ec", "network": {"id": "d946b94e-9827-4c22-af05-072fcaa3d0e6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.235", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ec39f2547f624ed0b240d084b949694a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe02d314-37", "ovs_interfaceid": "fe02d314-372f-43fb-8d7c-cfc33c9e74bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 765.880269] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:22:ec', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cb0e556a-0f69-4a5c-af62-ffc46edb8e63', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fe02d314-372f-43fb-8d7c-cfc33c9e74bd', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 765.889176] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Creating folder: Project (2141e7e1a0884ed1b512cf12e0d01e55). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 765.890073] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b4b0473b-6ed8-426e-a8e9-ecdc2f36f1b9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.900771] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Created folder: Project (2141e7e1a0884ed1b512cf12e0d01e55) in parent group-v264198. [ 765.901139] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Creating folder: Instances. Parent ref: group-v264215. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 765.901440] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-180978be-8502-4e42-a41c-478abd30ea93 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.910248] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Created folder: Instances in parent group-v264215. [ 765.910474] env[62600]: DEBUG oslo.service.loopingcall [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 765.910653] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 765.910841] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-08d84aae-ebbd-4e40-93e4-3e70f52be4be {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.932544] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 765.932544] env[62600]: value = "task-1222248" [ 765.932544] env[62600]: _type = "Task" [ 765.932544] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.940852] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222248, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.950751] env[62600]: DEBUG nova.compute.manager [req-162cfc3b-4f9a-494b-b941-0c9c084003d0 req-e996f93c-c873-40d2-9b2e-5be375239e11 service nova] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Received event network-changed-fe02d314-372f-43fb-8d7c-cfc33c9e74bd {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 765.950751] env[62600]: DEBUG nova.compute.manager [req-162cfc3b-4f9a-494b-b941-0c9c084003d0 req-e996f93c-c873-40d2-9b2e-5be375239e11 service nova] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Refreshing instance network info cache due to event network-changed-fe02d314-372f-43fb-8d7c-cfc33c9e74bd. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 765.950751] env[62600]: DEBUG oslo_concurrency.lockutils [req-162cfc3b-4f9a-494b-b941-0c9c084003d0 req-e996f93c-c873-40d2-9b2e-5be375239e11 service nova] Acquiring lock "refresh_cache-de4f9304-3357-4eaa-9c94-fe28bc554086" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.950751] env[62600]: DEBUG oslo_concurrency.lockutils [req-162cfc3b-4f9a-494b-b941-0c9c084003d0 req-e996f93c-c873-40d2-9b2e-5be375239e11 service nova] Acquired lock "refresh_cache-de4f9304-3357-4eaa-9c94-fe28bc554086" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.950751] env[62600]: DEBUG nova.network.neutron [req-162cfc3b-4f9a-494b-b941-0c9c084003d0 req-e996f93c-c873-40d2-9b2e-5be375239e11 service nova] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Refreshing network info cache for port fe02d314-372f-43fb-8d7c-cfc33c9e74bd {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 766.001628] env[62600]: DEBUG oslo_concurrency.lockutils [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.945s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.002254] env[62600]: ERROR nova.compute.manager [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e41ea304-1712-47b9-b8e8-e4faa3600e83, please check neutron logs for more information. [ 766.002254] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] Traceback (most recent call last): [ 766.002254] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 766.002254] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] self.driver.spawn(context, instance, image_meta, [ 766.002254] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 766.002254] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] self._vmops.spawn(context, instance, image_meta, injected_files, [ 766.002254] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 766.002254] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] vm_ref = self.build_virtual_machine(instance, [ 766.002254] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 766.002254] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] vif_infos = vmwarevif.get_vif_info(self._session, [ 766.002254] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 766.003447] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] for vif in network_info: [ 766.003447] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 766.003447] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] return self._sync_wrapper(fn, *args, **kwargs) [ 766.003447] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 766.003447] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] self.wait() [ 766.003447] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 766.003447] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] self[:] = self._gt.wait() [ 766.003447] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 766.003447] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] return self._exit_event.wait() [ 766.003447] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 766.003447] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] current.throw(*self._exc) [ 766.003447] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.003447] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] result = function(*args, **kwargs) [ 766.004376] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 766.004376] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] return func(*args, **kwargs) [ 766.004376] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 766.004376] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] raise e [ 766.004376] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.004376] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] nwinfo = self.network_api.allocate_for_instance( [ 766.004376] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 766.004376] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] created_port_ids = self._update_ports_for_instance( [ 766.004376] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 766.004376] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] with excutils.save_and_reraise_exception(): [ 766.004376] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.004376] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] self.force_reraise() [ 766.004376] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.005587] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] raise self.value [ 766.005587] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 766.005587] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] updated_port = self._update_port( [ 766.005587] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.005587] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] _ensure_no_port_binding_failure(port) [ 766.005587] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.005587] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] raise exception.PortBindingFailed(port_id=port['id']) [ 766.005587] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] nova.exception.PortBindingFailed: Binding failed for port e41ea304-1712-47b9-b8e8-e4faa3600e83, please check neutron logs for more information. [ 766.005587] env[62600]: ERROR nova.compute.manager [instance: 606becc8-ff43-469c-aca6-c35d67614960] [ 766.005587] env[62600]: DEBUG nova.compute.utils [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Binding failed for port e41ea304-1712-47b9-b8e8-e4faa3600e83, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 766.005915] env[62600]: DEBUG oslo_concurrency.lockutils [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.098s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.006640] env[62600]: INFO nova.compute.claims [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 766.009734] env[62600]: DEBUG nova.compute.manager [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Build of instance 606becc8-ff43-469c-aca6-c35d67614960 was re-scheduled: Binding failed for port e41ea304-1712-47b9-b8e8-e4faa3600e83, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 766.010603] env[62600]: DEBUG nova.compute.manager [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 766.010673] env[62600]: DEBUG oslo_concurrency.lockutils [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Acquiring lock "refresh_cache-606becc8-ff43-469c-aca6-c35d67614960" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.010928] env[62600]: DEBUG oslo_concurrency.lockutils [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Acquired lock "refresh_cache-606becc8-ff43-469c-aca6-c35d67614960" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.011009] env[62600]: DEBUG nova.network.neutron [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 766.445602] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222248, 'name': CreateVM_Task, 'duration_secs': 0.309021} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.445834] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 766.455879] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.456081] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.456416] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 766.456667] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71639a93-a772-4053-ad31-ca4d552e585e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.462412] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 766.462412] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]523b75b3-eafd-8c92-3508-035016f4b5fd" [ 766.462412] env[62600]: _type = "Task" [ 766.462412] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.471494] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]523b75b3-eafd-8c92-3508-035016f4b5fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.539688] env[62600]: DEBUG nova.network.neutron [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 766.713457] env[62600]: DEBUG nova.network.neutron [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.889288] env[62600]: DEBUG nova.network.neutron [req-162cfc3b-4f9a-494b-b941-0c9c084003d0 req-e996f93c-c873-40d2-9b2e-5be375239e11 service nova] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Updated VIF entry in instance network info cache for port fe02d314-372f-43fb-8d7c-cfc33c9e74bd. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 766.890504] env[62600]: DEBUG nova.network.neutron [req-162cfc3b-4f9a-494b-b941-0c9c084003d0 req-e996f93c-c873-40d2-9b2e-5be375239e11 service nova] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Updating instance_info_cache with network_info: [{"id": "fe02d314-372f-43fb-8d7c-cfc33c9e74bd", "address": "fa:16:3e:86:22:ec", "network": {"id": "d946b94e-9827-4c22-af05-072fcaa3d0e6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.235", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ec39f2547f624ed0b240d084b949694a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe02d314-37", "ovs_interfaceid": "fe02d314-372f-43fb-8d7c-cfc33c9e74bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.917607] env[62600]: DEBUG nova.network.neutron [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Successfully updated port: 779d4e97-daa9-44ec-9ee0-91658a73bc02 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 766.972542] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]523b75b3-eafd-8c92-3508-035016f4b5fd, 'name': SearchDatastore_Task, 'duration_secs': 0.010109} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.972953] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.973212] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 766.973444] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.973588] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.973765] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 766.974026] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-afbb11f5-b291-4fff-bf0c-7fcb04dcf8b2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.981552] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 766.981728] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 766.982423] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99bbfffb-1b02-45a4-bf00-6447222224ef {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.987586] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 766.987586] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]524f0f73-7581-3f9e-6218-2523e9f6f1b1" [ 766.987586] env[62600]: _type = "Task" [ 766.987586] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.995042] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]524f0f73-7581-3f9e-6218-2523e9f6f1b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.004590] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.004781] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.126486] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Acquiring lock "ee7175b9-dbe6-4f90-bd2b-8829194dc6c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.126761] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Lock "ee7175b9-dbe6-4f90-bd2b-8829194dc6c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.220778] env[62600]: DEBUG oslo_concurrency.lockutils [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Releasing lock "refresh_cache-606becc8-ff43-469c-aca6-c35d67614960" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.220778] env[62600]: DEBUG nova.compute.manager [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 767.221990] env[62600]: DEBUG nova.compute.manager [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 767.222165] env[62600]: DEBUG nova.network.neutron [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 767.244871] env[62600]: DEBUG nova.network.neutron [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 767.341331] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0efabebd-61a2-40a8-a08d-b17c3af17386 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.348993] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a05f01d-590c-4003-afc3-22666c95d702 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.378670] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7eed31c-e482-4e02-a9d8-0aab97aef91b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.387100] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a24d6c25-e7fd-4a20-be05-bd3aadf48b7c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.400884] env[62600]: DEBUG oslo_concurrency.lockutils [req-162cfc3b-4f9a-494b-b941-0c9c084003d0 req-e996f93c-c873-40d2-9b2e-5be375239e11 service nova] Releasing lock "refresh_cache-de4f9304-3357-4eaa-9c94-fe28bc554086" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.401486] env[62600]: DEBUG nova.compute.provider_tree [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.419663] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Acquiring lock "refresh_cache-92cf1acb-b471-4add-a73f-7e9e94fbaaa4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.419663] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Acquired lock "refresh_cache-92cf1acb-b471-4add-a73f-7e9e94fbaaa4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.419663] env[62600]: DEBUG nova.network.neutron [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 767.497801] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]524f0f73-7581-3f9e-6218-2523e9f6f1b1, 'name': SearchDatastore_Task, 'duration_secs': 0.008333} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.499204] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-561d5d2b-5158-4e42-ac33-e0a0cf0987f5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.503587] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 767.503587] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521f24d7-6b60-29e7-0b46-bfee5f553290" [ 767.503587] env[62600]: _type = "Task" [ 767.503587] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.511797] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521f24d7-6b60-29e7-0b46-bfee5f553290, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.511797] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 767.511797] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Starting heal instance info cache {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 767.511949] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Rebuilding the list of instances to heal {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 767.747741] env[62600]: DEBUG nova.network.neutron [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.819383] env[62600]: DEBUG nova.compute.manager [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 767.819943] env[62600]: DEBUG nova.virt.hardware [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 767.820190] env[62600]: DEBUG nova.virt.hardware [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 767.820358] env[62600]: DEBUG nova.virt.hardware [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 767.821521] env[62600]: DEBUG nova.virt.hardware [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 767.821937] env[62600]: DEBUG nova.virt.hardware [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 767.822042] env[62600]: DEBUG nova.virt.hardware [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 767.822259] env[62600]: DEBUG nova.virt.hardware [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 767.822469] env[62600]: DEBUG nova.virt.hardware [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 767.823396] env[62600]: DEBUG nova.virt.hardware [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 767.823396] env[62600]: DEBUG nova.virt.hardware [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 767.823396] env[62600]: DEBUG nova.virt.hardware [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 767.824556] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51275d0c-0755-407b-b929-cca6429c5af6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.833995] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe58303-4f59-4147-8408-b9e1f19b2767 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.904737] env[62600]: DEBUG nova.scheduler.client.report [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 767.967346] env[62600]: DEBUG nova.network.neutron [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 768.006311] env[62600]: DEBUG nova.compute.manager [req-7ac2f4ad-83a7-46ab-8d3c-dafaab66bc6e req-f1dbf648-90b1-42b7-95c4-6c9cc0a33a97 service nova] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Received event network-vif-plugged-779d4e97-daa9-44ec-9ee0-91658a73bc02 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.006527] env[62600]: DEBUG oslo_concurrency.lockutils [req-7ac2f4ad-83a7-46ab-8d3c-dafaab66bc6e req-f1dbf648-90b1-42b7-95c4-6c9cc0a33a97 service nova] Acquiring lock "92cf1acb-b471-4add-a73f-7e9e94fbaaa4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.006885] env[62600]: DEBUG oslo_concurrency.lockutils [req-7ac2f4ad-83a7-46ab-8d3c-dafaab66bc6e req-f1dbf648-90b1-42b7-95c4-6c9cc0a33a97 service nova] Lock "92cf1acb-b471-4add-a73f-7e9e94fbaaa4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.006885] env[62600]: DEBUG oslo_concurrency.lockutils [req-7ac2f4ad-83a7-46ab-8d3c-dafaab66bc6e req-f1dbf648-90b1-42b7-95c4-6c9cc0a33a97 service nova] Lock "92cf1acb-b471-4add-a73f-7e9e94fbaaa4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.007571] env[62600]: DEBUG nova.compute.manager [req-7ac2f4ad-83a7-46ab-8d3c-dafaab66bc6e req-f1dbf648-90b1-42b7-95c4-6c9cc0a33a97 service nova] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] No waiting events found dispatching network-vif-plugged-779d4e97-daa9-44ec-9ee0-91658a73bc02 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 768.007809] env[62600]: WARNING nova.compute.manager [req-7ac2f4ad-83a7-46ab-8d3c-dafaab66bc6e req-f1dbf648-90b1-42b7-95c4-6c9cc0a33a97 service nova] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Received unexpected event network-vif-plugged-779d4e97-daa9-44ec-9ee0-91658a73bc02 for instance with vm_state building and task_state spawning. [ 768.008041] env[62600]: DEBUG nova.compute.manager [req-7ac2f4ad-83a7-46ab-8d3c-dafaab66bc6e req-f1dbf648-90b1-42b7-95c4-6c9cc0a33a97 service nova] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Received event network-changed-779d4e97-daa9-44ec-9ee0-91658a73bc02 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.008219] env[62600]: DEBUG nova.compute.manager [req-7ac2f4ad-83a7-46ab-8d3c-dafaab66bc6e req-f1dbf648-90b1-42b7-95c4-6c9cc0a33a97 service nova] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Refreshing instance network info cache due to event network-changed-779d4e97-daa9-44ec-9ee0-91658a73bc02. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 768.008380] env[62600]: DEBUG oslo_concurrency.lockutils [req-7ac2f4ad-83a7-46ab-8d3c-dafaab66bc6e req-f1dbf648-90b1-42b7-95c4-6c9cc0a33a97 service nova] Acquiring lock "refresh_cache-92cf1acb-b471-4add-a73f-7e9e94fbaaa4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.015268] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Skipping network cache update for instance because it is Building. {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 768.015421] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Skipping network cache update for instance because it is Building. {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 768.015552] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Skipping network cache update for instance because it is Building. {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 768.015677] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Skipping network cache update for instance because it is Building. {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 768.015801] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Didn't find any instances for network info cache update. {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 768.016633] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.017146] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.017311] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.017463] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.017605] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.017819] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.017930] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62600) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 768.018109] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.024785] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521f24d7-6b60-29e7-0b46-bfee5f553290, 'name': SearchDatastore_Task, 'duration_secs': 0.008954} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.025055] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.025369] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] de4f9304-3357-4eaa-9c94-fe28bc554086/de4f9304-3357-4eaa-9c94-fe28bc554086.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 768.025641] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b4248073-e86d-42b6-8891-cd800c0fa278 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.034038] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 768.034038] env[62600]: value = "task-1222249" [ 768.034038] env[62600]: _type = "Task" [ 768.034038] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.043802] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222249, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.114991] env[62600]: DEBUG nova.network.neutron [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Updating instance_info_cache with network_info: [{"id": "779d4e97-daa9-44ec-9ee0-91658a73bc02", "address": "fa:16:3e:d5:8d:e6", "network": {"id": "29b27b28-57ba-466c-a0b5-38469adbe1d5", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1563603335-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c5a7585f2a4d7888f3240992a68955", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap779d4e97-da", "ovs_interfaceid": "779d4e97-daa9-44ec-9ee0-91658a73bc02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.255071] env[62600]: INFO nova.compute.manager [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] [instance: 606becc8-ff43-469c-aca6-c35d67614960] Took 1.03 seconds to deallocate network for instance. [ 768.411156] env[62600]: DEBUG oslo_concurrency.lockutils [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.407s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.411998] env[62600]: DEBUG nova.compute.manager [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 768.414224] env[62600]: DEBUG oslo_concurrency.lockutils [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.575s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.415697] env[62600]: INFO nova.compute.claims [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 768.520851] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.544027] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222249, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.44202} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.544177] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] de4f9304-3357-4eaa-9c94-fe28bc554086/de4f9304-3357-4eaa-9c94-fe28bc554086.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 768.544391] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 768.545077] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-27ad538a-74f6-4675-948c-6efc23c0d055 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.551575] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 768.551575] env[62600]: value = "task-1222250" [ 768.551575] env[62600]: _type = "Task" [ 768.551575] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.559627] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222250, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.618105] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Releasing lock "refresh_cache-92cf1acb-b471-4add-a73f-7e9e94fbaaa4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.618472] env[62600]: DEBUG nova.compute.manager [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Instance network_info: |[{"id": "779d4e97-daa9-44ec-9ee0-91658a73bc02", "address": "fa:16:3e:d5:8d:e6", "network": {"id": "29b27b28-57ba-466c-a0b5-38469adbe1d5", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1563603335-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c5a7585f2a4d7888f3240992a68955", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap779d4e97-da", "ovs_interfaceid": "779d4e97-daa9-44ec-9ee0-91658a73bc02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 768.618800] env[62600]: DEBUG oslo_concurrency.lockutils [req-7ac2f4ad-83a7-46ab-8d3c-dafaab66bc6e req-f1dbf648-90b1-42b7-95c4-6c9cc0a33a97 service nova] Acquired lock "refresh_cache-92cf1acb-b471-4add-a73f-7e9e94fbaaa4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.619062] env[62600]: DEBUG nova.network.neutron [req-7ac2f4ad-83a7-46ab-8d3c-dafaab66bc6e req-f1dbf648-90b1-42b7-95c4-6c9cc0a33a97 service nova] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Refreshing network info cache for port 779d4e97-daa9-44ec-9ee0-91658a73bc02 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 768.620559] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d5:8d:e6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c492f5cc-7ae0-4cab-823c-0d5dd8c60b26', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '779d4e97-daa9-44ec-9ee0-91658a73bc02', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 768.629149] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Creating folder: Project (49c5a7585f2a4d7888f3240992a68955). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 768.630410] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cbb358f4-653a-4aee-9e33-94ca834a3313 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.644722] env[62600]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 768.644948] env[62600]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62600) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 768.645348] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Folder already exists: Project (49c5a7585f2a4d7888f3240992a68955). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 768.645738] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Creating folder: Instances. Parent ref: group-v264208. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 768.645939] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ab98d228-fd95-4cad-824a-c64bb0ee75f6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.655803] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Created folder: Instances in parent group-v264208. [ 768.656062] env[62600]: DEBUG oslo.service.loopingcall [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 768.656275] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 768.656508] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-984a278b-7707-4889-896f-9685f5ed2e8b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.676594] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 768.676594] env[62600]: value = "task-1222253" [ 768.676594] env[62600]: _type = "Task" [ 768.676594] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.688825] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222253, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.920648] env[62600]: DEBUG nova.compute.utils [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 768.924604] env[62600]: DEBUG nova.compute.manager [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 768.924785] env[62600]: DEBUG nova.network.neutron [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 768.993822] env[62600]: DEBUG nova.policy [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd0443142c9543c88695e5cb99585f24', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5ab0065aa6e94abfae3bbd5cd6415c3b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 769.061937] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222250, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064655} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.061937] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 769.063072] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c8f5cff-4ff5-41f3-9f55-2d45ed126fa8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.084981] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] de4f9304-3357-4eaa-9c94-fe28bc554086/de4f9304-3357-4eaa-9c94-fe28bc554086.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 769.085528] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dfbd7b13-b342-45fa-afa7-a233f123aace {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.114999] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 769.114999] env[62600]: value = "task-1222254" [ 769.114999] env[62600]: _type = "Task" [ 769.114999] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.123959] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222254, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.190505] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222253, 'name': CreateVM_Task, 'duration_secs': 0.351353} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.190675] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 769.191364] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'mount_device': '/dev/sda', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264214', 'volume_id': '88d3934f-d0ae-4578-af24-9abb842074b3', 'name': 'volume-88d3934f-d0ae-4578-af24-9abb842074b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '92cf1acb-b471-4add-a73f-7e9e94fbaaa4', 'attached_at': '', 'detached_at': '', 'volume_id': '88d3934f-d0ae-4578-af24-9abb842074b3', 'serial': '88d3934f-d0ae-4578-af24-9abb842074b3'}, 'guest_format': None, 'attachment_id': 'f8573456-d219-47bb-87ae-26f22ee5429f', 'delete_on_termination': True, 'device_type': None, 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=62600) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 769.191598] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Root volume attach. Driver type: vmdk {{(pid=62600) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 769.192579] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03712ce5-7ace-46e4-9520-3608e559b7b8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.200703] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b1dbf2f-8f57-466a-a707-d11814eaa164 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.207218] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb90836-bcdb-4cba-a06a-84aab47a413d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.213590] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-4b5b200e-1229-462d-a6a7-15b8e4ea6fbd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.221038] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Waiting for the task: (returnval){ [ 769.221038] env[62600]: value = "task-1222255" [ 769.221038] env[62600]: _type = "Task" [ 769.221038] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.229702] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222255, 'name': RelocateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.286180] env[62600]: INFO nova.scheduler.client.report [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Deleted allocations for instance 606becc8-ff43-469c-aca6-c35d67614960 [ 769.411028] env[62600]: DEBUG nova.network.neutron [req-7ac2f4ad-83a7-46ab-8d3c-dafaab66bc6e req-f1dbf648-90b1-42b7-95c4-6c9cc0a33a97 service nova] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Updated VIF entry in instance network info cache for port 779d4e97-daa9-44ec-9ee0-91658a73bc02. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 769.411028] env[62600]: DEBUG nova.network.neutron [req-7ac2f4ad-83a7-46ab-8d3c-dafaab66bc6e req-f1dbf648-90b1-42b7-95c4-6c9cc0a33a97 service nova] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Updating instance_info_cache with network_info: [{"id": "779d4e97-daa9-44ec-9ee0-91658a73bc02", "address": "fa:16:3e:d5:8d:e6", "network": {"id": "29b27b28-57ba-466c-a0b5-38469adbe1d5", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1563603335-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c5a7585f2a4d7888f3240992a68955", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap779d4e97-da", "ovs_interfaceid": "779d4e97-daa9-44ec-9ee0-91658a73bc02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.426221] env[62600]: DEBUG nova.compute.manager [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 769.628759] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222254, 'name': ReconfigVM_Task, 'duration_secs': 0.312082} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.633623] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Reconfigured VM instance instance-0000002b to attach disk [datastore2] de4f9304-3357-4eaa-9c94-fe28bc554086/de4f9304-3357-4eaa-9c94-fe28bc554086.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 769.636177] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0823be84-63f9-461e-b535-007562e0f3e2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.645867] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 769.645867] env[62600]: value = "task-1222256" [ 769.645867] env[62600]: _type = "Task" [ 769.645867] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.657776] env[62600]: DEBUG nova.network.neutron [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Successfully created port: c5d84e4b-f903-4d3b-ad0f-4a596b324f41 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 769.670313] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222256, 'name': Rename_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.738179] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222255, 'name': RelocateVM_Task} progress is 40%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.797854] env[62600]: DEBUG oslo_concurrency.lockutils [None req-556fed11-756a-4d85-ad44-394767c177ba tempest-ServersNegativeTestJSON-2087328975 tempest-ServersNegativeTestJSON-2087328975-project-member] Lock "606becc8-ff43-469c-aca6-c35d67614960" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.370s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.840070] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f48d76a-68ad-4f08-8829-77e0be384216 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.849172] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e20aa8-ef29-4726-82e6-fb19188debb3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.883593] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e2317a-a6a1-4000-bb23-a95da787abba {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.894053] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-190d6584-0d8e-4206-a9a3-77aaba6306bf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.913016] env[62600]: DEBUG nova.compute.provider_tree [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 769.918401] env[62600]: DEBUG oslo_concurrency.lockutils [req-7ac2f4ad-83a7-46ab-8d3c-dafaab66bc6e req-f1dbf648-90b1-42b7-95c4-6c9cc0a33a97 service nova] Releasing lock "refresh_cache-92cf1acb-b471-4add-a73f-7e9e94fbaaa4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.160548] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222256, 'name': Rename_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.238071] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222255, 'name': RelocateVM_Task} progress is 54%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.303647] env[62600]: DEBUG nova.compute.manager [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 770.416824] env[62600]: DEBUG nova.scheduler.client.report [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 770.441433] env[62600]: DEBUG nova.compute.manager [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 770.469638] env[62600]: DEBUG nova.virt.hardware [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 770.469952] env[62600]: DEBUG nova.virt.hardware [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 770.470647] env[62600]: DEBUG nova.virt.hardware [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 770.470872] env[62600]: DEBUG nova.virt.hardware [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 770.471048] env[62600]: DEBUG nova.virt.hardware [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 770.471214] env[62600]: DEBUG nova.virt.hardware [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 770.471427] env[62600]: DEBUG nova.virt.hardware [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 770.471588] env[62600]: DEBUG nova.virt.hardware [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 770.471758] env[62600]: DEBUG nova.virt.hardware [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 770.471919] env[62600]: DEBUG nova.virt.hardware [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 770.472113] env[62600]: DEBUG nova.virt.hardware [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 770.473432] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b014fd8-7b97-46e3-8bdb-cf2121a24750 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.482835] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a375f2be-33b9-49b2-880a-eff6fe602de7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.661590] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222256, 'name': Rename_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.735136] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222255, 'name': RelocateVM_Task} progress is 67%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.836178] env[62600]: DEBUG oslo_concurrency.lockutils [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.922363] env[62600]: DEBUG oslo_concurrency.lockutils [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.508s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.923948] env[62600]: DEBUG nova.compute.manager [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 770.926246] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.780s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.159719] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222256, 'name': Rename_Task, 'duration_secs': 1.20208} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.159996] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 771.160251] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-82f615ec-35aa-46e1-9d36-98ac5acbeef3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.169033] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 771.169033] env[62600]: value = "task-1222257" [ 771.169033] env[62600]: _type = "Task" [ 771.169033] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.178667] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222257, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.241123] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222255, 'name': RelocateVM_Task} progress is 82%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.430291] env[62600]: DEBUG nova.compute.utils [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 771.430291] env[62600]: DEBUG nova.compute.manager [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 771.430291] env[62600]: DEBUG nova.network.neutron [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 771.573082] env[62600]: DEBUG nova.policy [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd0443142c9543c88695e5cb99585f24', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5ab0065aa6e94abfae3bbd5cd6415c3b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 771.687241] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222257, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.715255] env[62600]: DEBUG nova.network.neutron [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Successfully updated port: c5d84e4b-f903-4d3b-ad0f-4a596b324f41 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 771.723637] env[62600]: DEBUG nova.compute.manager [req-dc1a1b6c-02a3-45b4-b966-b755bdeba999 req-55066c5e-c47e-440a-b040-5258837e0cbf service nova] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Received event network-vif-plugged-c5d84e4b-f903-4d3b-ad0f-4a596b324f41 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 771.723637] env[62600]: DEBUG oslo_concurrency.lockutils [req-dc1a1b6c-02a3-45b4-b966-b755bdeba999 req-55066c5e-c47e-440a-b040-5258837e0cbf service nova] Acquiring lock "efff4d72-4673-4a8d-9a81-be9d700ff881-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.723637] env[62600]: DEBUG oslo_concurrency.lockutils [req-dc1a1b6c-02a3-45b4-b966-b755bdeba999 req-55066c5e-c47e-440a-b040-5258837e0cbf service nova] Lock "efff4d72-4673-4a8d-9a81-be9d700ff881-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.723637] env[62600]: DEBUG oslo_concurrency.lockutils [req-dc1a1b6c-02a3-45b4-b966-b755bdeba999 req-55066c5e-c47e-440a-b040-5258837e0cbf service nova] Lock "efff4d72-4673-4a8d-9a81-be9d700ff881-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.723637] env[62600]: DEBUG nova.compute.manager [req-dc1a1b6c-02a3-45b4-b966-b755bdeba999 req-55066c5e-c47e-440a-b040-5258837e0cbf service nova] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] No waiting events found dispatching network-vif-plugged-c5d84e4b-f903-4d3b-ad0f-4a596b324f41 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 771.723842] env[62600]: WARNING nova.compute.manager [req-dc1a1b6c-02a3-45b4-b966-b755bdeba999 req-55066c5e-c47e-440a-b040-5258837e0cbf service nova] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Received unexpected event network-vif-plugged-c5d84e4b-f903-4d3b-ad0f-4a596b324f41 for instance with vm_state building and task_state spawning. [ 771.750741] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222255, 'name': RelocateVM_Task} progress is 97%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.867771] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d330c8eb-007e-4296-9005-aab656dcc3d7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.879497] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a2a13e9-1a73-40ff-ad44-0cbc4247cd65 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.916453] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2c4eca-e7d9-4e9b-819f-d9b54df1ba7f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.924709] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e651aeaa-63ea-41d1-aff5-288b75107ae6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.940139] env[62600]: DEBUG nova.compute.manager [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 771.942545] env[62600]: DEBUG nova.compute.provider_tree [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.061770] env[62600]: DEBUG nova.network.neutron [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Successfully created port: 6a3323e4-5bec-4c2d-8c89-285faea1dd06 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 772.081867] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "b967fb11-e70e-4e17-b769-38da581bd83b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.083896] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "b967fb11-e70e-4e17-b769-38da581bd83b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.183275] env[62600]: DEBUG oslo_vmware.api [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222257, 'name': PowerOnVM_Task, 'duration_secs': 0.966134} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.183529] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 772.183733] env[62600]: INFO nova.compute.manager [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Took 8.52 seconds to spawn the instance on the hypervisor. [ 772.183918] env[62600]: DEBUG nova.compute.manager [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 772.184831] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4c54db8-3660-4843-9c62-ed5bf6f27e42 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.218560] env[62600]: DEBUG oslo_concurrency.lockutils [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquiring lock "refresh_cache-efff4d72-4673-4a8d-9a81-be9d700ff881" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.218757] env[62600]: DEBUG oslo_concurrency.lockutils [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquired lock "refresh_cache-efff4d72-4673-4a8d-9a81-be9d700ff881" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.218910] env[62600]: DEBUG nova.network.neutron [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 772.240866] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222255, 'name': RelocateVM_Task} progress is 97%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.452971] env[62600]: DEBUG nova.scheduler.client.report [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 772.710024] env[62600]: INFO nova.compute.manager [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Took 27.17 seconds to build instance. [ 772.742382] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222255, 'name': RelocateVM_Task} progress is 97%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.768174] env[62600]: DEBUG nova.network.neutron [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 772.924122] env[62600]: DEBUG nova.network.neutron [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Updating instance_info_cache with network_info: [{"id": "c5d84e4b-f903-4d3b-ad0f-4a596b324f41", "address": "fa:16:3e:cb:af:aa", "network": {"id": "d946b94e-9827-4c22-af05-072fcaa3d0e6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.209", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ec39f2547f624ed0b240d084b949694a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5d84e4b-f9", "ovs_interfaceid": "c5d84e4b-f903-4d3b-ad0f-4a596b324f41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.952353] env[62600]: DEBUG nova.compute.manager [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 772.960411] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.033s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.960411] env[62600]: ERROR nova.compute.manager [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 42193f90-0f3e-4b54-8eb0-f5ec95237a03, please check neutron logs for more information. [ 772.960411] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Traceback (most recent call last): [ 772.960411] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 772.960411] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] self.driver.spawn(context, instance, image_meta, [ 772.960411] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 772.960411] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 772.960411] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 772.960411] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] vm_ref = self.build_virtual_machine(instance, [ 772.960772] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 772.960772] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] vif_infos = vmwarevif.get_vif_info(self._session, [ 772.960772] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 772.960772] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] for vif in network_info: [ 772.960772] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 772.960772] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] return self._sync_wrapper(fn, *args, **kwargs) [ 772.960772] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 772.960772] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] self.wait() [ 772.960772] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 772.960772] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] self[:] = self._gt.wait() [ 772.960772] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 772.960772] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] return self._exit_event.wait() [ 772.960772] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 772.961159] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] current.throw(*self._exc) [ 772.961159] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 772.961159] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] result = function(*args, **kwargs) [ 772.961159] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 772.961159] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] return func(*args, **kwargs) [ 772.961159] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 772.961159] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] raise e [ 772.961159] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 772.961159] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] nwinfo = self.network_api.allocate_for_instance( [ 772.961159] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 772.961159] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] created_port_ids = self._update_ports_for_instance( [ 772.961159] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 772.961159] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] with excutils.save_and_reraise_exception(): [ 772.961514] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 772.961514] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] self.force_reraise() [ 772.961514] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 772.961514] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] raise self.value [ 772.961514] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 772.961514] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] updated_port = self._update_port( [ 772.961514] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 772.961514] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] _ensure_no_port_binding_failure(port) [ 772.961514] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 772.961514] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] raise exception.PortBindingFailed(port_id=port['id']) [ 772.961514] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] nova.exception.PortBindingFailed: Binding failed for port 42193f90-0f3e-4b54-8eb0-f5ec95237a03, please check neutron logs for more information. [ 772.961514] env[62600]: ERROR nova.compute.manager [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] [ 772.961900] env[62600]: DEBUG nova.compute.utils [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Binding failed for port 42193f90-0f3e-4b54-8eb0-f5ec95237a03, please check neutron logs for more information. {{(pid=62600) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 772.962576] env[62600]: DEBUG oslo_concurrency.lockutils [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.687s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.965270] env[62600]: INFO nova.compute.claims [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 772.968623] env[62600]: DEBUG nova.compute.manager [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Build of instance 014899cb-db4c-4444-abe1-b0129c0d7db9 was re-scheduled: Binding failed for port 42193f90-0f3e-4b54-8eb0-f5ec95237a03, please check neutron logs for more information. {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 772.969314] env[62600]: DEBUG nova.compute.manager [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Unplugging VIFs for instance {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 772.969721] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Acquiring lock "refresh_cache-014899cb-db4c-4444-abe1-b0129c0d7db9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.969721] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Acquired lock "refresh_cache-014899cb-db4c-4444-abe1-b0129c0d7db9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.969849] env[62600]: DEBUG nova.network.neutron [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 772.983396] env[62600]: DEBUG nova.virt.hardware [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 772.983879] env[62600]: DEBUG nova.virt.hardware [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 772.983985] env[62600]: DEBUG nova.virt.hardware [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 772.984245] env[62600]: DEBUG nova.virt.hardware [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 772.984828] env[62600]: DEBUG nova.virt.hardware [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 772.984828] env[62600]: DEBUG nova.virt.hardware [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 772.985471] env[62600]: DEBUG nova.virt.hardware [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 772.985471] env[62600]: DEBUG nova.virt.hardware [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 772.985693] env[62600]: DEBUG nova.virt.hardware [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 772.986035] env[62600]: DEBUG nova.virt.hardware [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 772.986267] env[62600]: DEBUG nova.virt.hardware [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 772.987430] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de530813-7e98-48c8-813d-d3fe4a9adec0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.996992] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd83e07a-7e16-4037-ae57-91f2baf60f90 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.064922] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "d66c404d-8fd1-4fb7-a3b9-f21854c7e735" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.065238] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "d66c404d-8fd1-4fb7-a3b9-f21854c7e735" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.215380] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a0e5f825-bd82-4672-af22-4e78ea81c482 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "de4f9304-3357-4eaa-9c94-fe28bc554086" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.856s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.236502] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222255, 'name': RelocateVM_Task} progress is 98%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.427315] env[62600]: DEBUG oslo_concurrency.lockutils [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Releasing lock "refresh_cache-efff4d72-4673-4a8d-9a81-be9d700ff881" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.427655] env[62600]: DEBUG nova.compute.manager [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Instance network_info: |[{"id": "c5d84e4b-f903-4d3b-ad0f-4a596b324f41", "address": "fa:16:3e:cb:af:aa", "network": {"id": "d946b94e-9827-4c22-af05-072fcaa3d0e6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.209", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ec39f2547f624ed0b240d084b949694a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5d84e4b-f9", "ovs_interfaceid": "c5d84e4b-f903-4d3b-ad0f-4a596b324f41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 773.428151] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cb:af:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cb0e556a-0f69-4a5c-af62-ffc46edb8e63', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c5d84e4b-f903-4d3b-ad0f-4a596b324f41', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 773.436169] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Creating folder: Project (5ab0065aa6e94abfae3bbd5cd6415c3b). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 773.436486] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c900b1c4-9cec-4c04-a307-72c510e25e3b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.447386] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Created folder: Project (5ab0065aa6e94abfae3bbd5cd6415c3b) in parent group-v264198. [ 773.447527] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Creating folder: Instances. Parent ref: group-v264220. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 773.447770] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-238c5d7d-bd1f-4685-8843-f0fa0dd0137a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.457096] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Created folder: Instances in parent group-v264220. [ 773.457359] env[62600]: DEBUG oslo.service.loopingcall [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 773.457555] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 773.457758] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e26b50a5-3352-411d-a0dc-02d1f3fb7ae2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.480456] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 773.480456] env[62600]: value = "task-1222260" [ 773.480456] env[62600]: _type = "Task" [ 773.480456] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.489721] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222260, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.490543] env[62600]: DEBUG nova.network.neutron [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 773.572621] env[62600]: DEBUG nova.network.neutron [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.720222] env[62600]: DEBUG nova.compute.manager [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 773.740614] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222255, 'name': RelocateVM_Task, 'duration_secs': 4.098109} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.740614] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Volume attach. Driver type: vmdk {{(pid=62600) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 773.740614] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264214', 'volume_id': '88d3934f-d0ae-4578-af24-9abb842074b3', 'name': 'volume-88d3934f-d0ae-4578-af24-9abb842074b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '92cf1acb-b471-4add-a73f-7e9e94fbaaa4', 'attached_at': '', 'detached_at': '', 'volume_id': '88d3934f-d0ae-4578-af24-9abb842074b3', 'serial': '88d3934f-d0ae-4578-af24-9abb842074b3'} {{(pid=62600) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 773.740614] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5cefdca-2929-4210-a56a-c414d064b2cf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.755689] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecae02f5-fe23-4f7b-822c-8250c55a45d9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.760730] env[62600]: DEBUG nova.compute.manager [req-74448599-238a-4081-bcac-80969615a81c req-0b2721c7-05ef-4214-b41c-a2eaf7b63c19 service nova] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Received event network-changed-c5d84e4b-f903-4d3b-ad0f-4a596b324f41 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.761262] env[62600]: DEBUG nova.compute.manager [req-74448599-238a-4081-bcac-80969615a81c req-0b2721c7-05ef-4214-b41c-a2eaf7b63c19 service nova] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Refreshing instance network info cache due to event network-changed-c5d84e4b-f903-4d3b-ad0f-4a596b324f41. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 773.761498] env[62600]: DEBUG oslo_concurrency.lockutils [req-74448599-238a-4081-bcac-80969615a81c req-0b2721c7-05ef-4214-b41c-a2eaf7b63c19 service nova] Acquiring lock "refresh_cache-efff4d72-4673-4a8d-9a81-be9d700ff881" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.761643] env[62600]: DEBUG oslo_concurrency.lockutils [req-74448599-238a-4081-bcac-80969615a81c req-0b2721c7-05ef-4214-b41c-a2eaf7b63c19 service nova] Acquired lock "refresh_cache-efff4d72-4673-4a8d-9a81-be9d700ff881" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.761803] env[62600]: DEBUG nova.network.neutron [req-74448599-238a-4081-bcac-80969615a81c req-0b2721c7-05ef-4214-b41c-a2eaf7b63c19 service nova] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Refreshing network info cache for port c5d84e4b-f903-4d3b-ad0f-4a596b324f41 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 773.784914] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] volume-88d3934f-d0ae-4578-af24-9abb842074b3/volume-88d3934f-d0ae-4578-af24-9abb842074b3.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 773.785782] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a736666-4c5d-4e21-915f-e81cf9661d65 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.806115] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Waiting for the task: (returnval){ [ 773.806115] env[62600]: value = "task-1222261" [ 773.806115] env[62600]: _type = "Task" [ 773.806115] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.814696] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222261, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.989959] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222260, 'name': CreateVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.061461] env[62600]: DEBUG nova.compute.manager [req-d24ed197-6f79-4e17-8c82-0905f65bdf52 req-1c3b7018-545b-4be9-a31e-e2c077f606df service nova] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Received event network-vif-plugged-6a3323e4-5bec-4c2d-8c89-285faea1dd06 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 774.061683] env[62600]: DEBUG oslo_concurrency.lockutils [req-d24ed197-6f79-4e17-8c82-0905f65bdf52 req-1c3b7018-545b-4be9-a31e-e2c077f606df service nova] Acquiring lock "a358a3cb-deda-419a-aa3c-ce7aeb534240-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.061870] env[62600]: DEBUG oslo_concurrency.lockutils [req-d24ed197-6f79-4e17-8c82-0905f65bdf52 req-1c3b7018-545b-4be9-a31e-e2c077f606df service nova] Lock "a358a3cb-deda-419a-aa3c-ce7aeb534240-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.062456] env[62600]: DEBUG oslo_concurrency.lockutils [req-d24ed197-6f79-4e17-8c82-0905f65bdf52 req-1c3b7018-545b-4be9-a31e-e2c077f606df service nova] Lock "a358a3cb-deda-419a-aa3c-ce7aeb534240-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.062827] env[62600]: DEBUG nova.compute.manager [req-d24ed197-6f79-4e17-8c82-0905f65bdf52 req-1c3b7018-545b-4be9-a31e-e2c077f606df service nova] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] No waiting events found dispatching network-vif-plugged-6a3323e4-5bec-4c2d-8c89-285faea1dd06 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 774.062827] env[62600]: WARNING nova.compute.manager [req-d24ed197-6f79-4e17-8c82-0905f65bdf52 req-1c3b7018-545b-4be9-a31e-e2c077f606df service nova] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Received unexpected event network-vif-plugged-6a3323e4-5bec-4c2d-8c89-285faea1dd06 for instance with vm_state building and task_state spawning. [ 774.075708] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Releasing lock "refresh_cache-014899cb-db4c-4444-abe1-b0129c0d7db9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.076069] env[62600]: DEBUG nova.compute.manager [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62600) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 774.077126] env[62600]: DEBUG nova.compute.manager [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 774.077126] env[62600]: DEBUG nova.network.neutron [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 774.099917] env[62600]: DEBUG nova.network.neutron [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 774.240561] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.289799] env[62600]: DEBUG nova.network.neutron [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Successfully updated port: 6a3323e4-5bec-4c2d-8c89-285faea1dd06 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 774.323927] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222261, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.381488] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba5aa03-7de2-45c9-b044-ad7adec319c5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.391081] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d81cd1-48e6-4c0f-be43-00400a6ae63b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.426528] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1b3fb83-13e3-4f4b-af5c-6957144b91d6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.436799] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f688314a-0a47-4991-8e2b-93036e2eb000 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.448055] env[62600]: DEBUG nova.compute.provider_tree [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 774.490419] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222260, 'name': CreateVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.536093] env[62600]: DEBUG nova.network.neutron [req-74448599-238a-4081-bcac-80969615a81c req-0b2721c7-05ef-4214-b41c-a2eaf7b63c19 service nova] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Updated VIF entry in instance network info cache for port c5d84e4b-f903-4d3b-ad0f-4a596b324f41. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 774.536478] env[62600]: DEBUG nova.network.neutron [req-74448599-238a-4081-bcac-80969615a81c req-0b2721c7-05ef-4214-b41c-a2eaf7b63c19 service nova] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Updating instance_info_cache with network_info: [{"id": "c5d84e4b-f903-4d3b-ad0f-4a596b324f41", "address": "fa:16:3e:cb:af:aa", "network": {"id": "d946b94e-9827-4c22-af05-072fcaa3d0e6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.209", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ec39f2547f624ed0b240d084b949694a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5d84e4b-f9", "ovs_interfaceid": "c5d84e4b-f903-4d3b-ad0f-4a596b324f41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.568838] env[62600]: DEBUG nova.compute.manager [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Stashing vm_state: active {{(pid=62600) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 774.603421] env[62600]: DEBUG nova.network.neutron [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.794523] env[62600]: DEBUG oslo_concurrency.lockutils [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquiring lock "refresh_cache-a358a3cb-deda-419a-aa3c-ce7aeb534240" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.794692] env[62600]: DEBUG oslo_concurrency.lockutils [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquired lock "refresh_cache-a358a3cb-deda-419a-aa3c-ce7aeb534240" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.794840] env[62600]: DEBUG nova.network.neutron [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 774.817690] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222261, 'name': ReconfigVM_Task, 'duration_secs': 0.55115} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.818072] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Reconfigured VM instance instance-0000002c to attach disk [datastore2] volume-88d3934f-d0ae-4578-af24-9abb842074b3/volume-88d3934f-d0ae-4578-af24-9abb842074b3.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 774.823161] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e66094f5-2e15-408c-a0da-8de833b9ed2f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.838356] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Waiting for the task: (returnval){ [ 774.838356] env[62600]: value = "task-1222262" [ 774.838356] env[62600]: _type = "Task" [ 774.838356] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.846315] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222262, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.950853] env[62600]: DEBUG nova.scheduler.client.report [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 774.994416] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222260, 'name': CreateVM_Task, 'duration_secs': 1.48219} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.994599] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 774.995652] env[62600]: DEBUG oslo_concurrency.lockutils [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.995937] env[62600]: DEBUG oslo_concurrency.lockutils [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.996410] env[62600]: DEBUG oslo_concurrency.lockutils [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 774.996743] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b00c0fc-f15f-4357-8e2d-362ba842a487 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.002464] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 775.002464] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f44173-2b30-4390-a2ca-80a709db8977" [ 775.002464] env[62600]: _type = "Task" [ 775.002464] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.013357] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f44173-2b30-4390-a2ca-80a709db8977, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.039659] env[62600]: DEBUG oslo_concurrency.lockutils [req-74448599-238a-4081-bcac-80969615a81c req-0b2721c7-05ef-4214-b41c-a2eaf7b63c19 service nova] Releasing lock "refresh_cache-efff4d72-4673-4a8d-9a81-be9d700ff881" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.088543] env[62600]: DEBUG oslo_concurrency.lockutils [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.107121] env[62600]: INFO nova.compute.manager [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] [instance: 014899cb-db4c-4444-abe1-b0129c0d7db9] Took 1.03 seconds to deallocate network for instance. [ 775.325010] env[62600]: DEBUG nova.network.neutron [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 775.349362] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222262, 'name': ReconfigVM_Task, 'duration_secs': 0.124483} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.349747] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264214', 'volume_id': '88d3934f-d0ae-4578-af24-9abb842074b3', 'name': 'volume-88d3934f-d0ae-4578-af24-9abb842074b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '92cf1acb-b471-4add-a73f-7e9e94fbaaa4', 'attached_at': '', 'detached_at': '', 'volume_id': '88d3934f-d0ae-4578-af24-9abb842074b3', 'serial': '88d3934f-d0ae-4578-af24-9abb842074b3'} {{(pid=62600) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 775.350384] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-02458c79-3364-4cc0-857f-12da21596fa0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.355931] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Waiting for the task: (returnval){ [ 775.355931] env[62600]: value = "task-1222263" [ 775.355931] env[62600]: _type = "Task" [ 775.355931] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.363136] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222263, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.456688] env[62600]: DEBUG oslo_concurrency.lockutils [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.494s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.457241] env[62600]: DEBUG nova.compute.manager [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 775.460742] env[62600]: DEBUG oslo_concurrency.lockutils [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.277s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.462537] env[62600]: INFO nova.compute.claims [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 775.477334] env[62600]: DEBUG nova.network.neutron [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Updating instance_info_cache with network_info: [{"id": "6a3323e4-5bec-4c2d-8c89-285faea1dd06", "address": "fa:16:3e:8e:12:44", "network": {"id": "d946b94e-9827-4c22-af05-072fcaa3d0e6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.176", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ec39f2547f624ed0b240d084b949694a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a3323e4-5b", "ovs_interfaceid": "6a3323e4-5bec-4c2d-8c89-285faea1dd06", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.513562] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f44173-2b30-4390-a2ca-80a709db8977, 'name': SearchDatastore_Task, 'duration_secs': 0.010234} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.513562] env[62600]: DEBUG oslo_concurrency.lockutils [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.513735] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 775.513951] env[62600]: DEBUG oslo_concurrency.lockutils [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.514108] env[62600]: DEBUG oslo_concurrency.lockutils [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.514291] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 775.514543] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d1697cc7-51a6-48f0-b0d3-6acfc8b11dfe {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.522803] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 775.522977] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 775.523687] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9dfc7fca-16d7-4bdd-827e-4b1e8c860a33 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.529042] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 775.529042] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520c7aca-615d-c581-62b8-83149925b791" [ 775.529042] env[62600]: _type = "Task" [ 775.529042] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.535934] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520c7aca-615d-c581-62b8-83149925b791, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.867222] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222263, 'name': Rename_Task, 'duration_secs': 0.133186} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.867509] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 775.867801] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-23534d6f-357f-4bbd-811b-b36d81007ec7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.873428] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Waiting for the task: (returnval){ [ 775.873428] env[62600]: value = "task-1222264" [ 775.873428] env[62600]: _type = "Task" [ 775.873428] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.881994] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222264, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.962162] env[62600]: DEBUG nova.compute.utils [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 775.963544] env[62600]: DEBUG nova.compute.manager [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 775.963712] env[62600]: DEBUG nova.network.neutron [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 775.981146] env[62600]: DEBUG oslo_concurrency.lockutils [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Releasing lock "refresh_cache-a358a3cb-deda-419a-aa3c-ce7aeb534240" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.981422] env[62600]: DEBUG nova.compute.manager [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Instance network_info: |[{"id": "6a3323e4-5bec-4c2d-8c89-285faea1dd06", "address": "fa:16:3e:8e:12:44", "network": {"id": "d946b94e-9827-4c22-af05-072fcaa3d0e6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.176", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ec39f2547f624ed0b240d084b949694a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a3323e4-5b", "ovs_interfaceid": "6a3323e4-5bec-4c2d-8c89-285faea1dd06", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 775.981799] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:12:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cb0e556a-0f69-4a5c-af62-ffc46edb8e63', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6a3323e4-5bec-4c2d-8c89-285faea1dd06', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 775.989192] env[62600]: DEBUG oslo.service.loopingcall [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 775.989601] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 775.989813] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7771c756-f4c5-4130-9c21-69f7e0d70adf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.005274] env[62600]: DEBUG nova.policy [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3ca871780304f5c88db2bcf163f2dea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4e28048248a44a6fb0113b3f84a8e8be', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 776.011673] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 776.011673] env[62600]: value = "task-1222265" [ 776.011673] env[62600]: _type = "Task" [ 776.011673] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.020620] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222265, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.036738] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520c7aca-615d-c581-62b8-83149925b791, 'name': SearchDatastore_Task, 'duration_secs': 0.008081} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.037542] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3effef38-e944-465a-a517-9d0e06d63118 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.042094] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 776.042094] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e33e37-638e-940a-df1a-3f047d9d5920" [ 776.042094] env[62600]: _type = "Task" [ 776.042094] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.048923] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e33e37-638e-940a-df1a-3f047d9d5920, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.085839] env[62600]: DEBUG nova.compute.manager [req-2c3a0c08-47ce-43c4-b146-737338f2e7ac req-509560fa-fb2c-4b03-af5c-083158448b1c service nova] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Received event network-changed-6a3323e4-5bec-4c2d-8c89-285faea1dd06 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 776.085839] env[62600]: DEBUG nova.compute.manager [req-2c3a0c08-47ce-43c4-b146-737338f2e7ac req-509560fa-fb2c-4b03-af5c-083158448b1c service nova] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Refreshing instance network info cache due to event network-changed-6a3323e4-5bec-4c2d-8c89-285faea1dd06. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 776.085839] env[62600]: DEBUG oslo_concurrency.lockutils [req-2c3a0c08-47ce-43c4-b146-737338f2e7ac req-509560fa-fb2c-4b03-af5c-083158448b1c service nova] Acquiring lock "refresh_cache-a358a3cb-deda-419a-aa3c-ce7aeb534240" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.086072] env[62600]: DEBUG oslo_concurrency.lockutils [req-2c3a0c08-47ce-43c4-b146-737338f2e7ac req-509560fa-fb2c-4b03-af5c-083158448b1c service nova] Acquired lock "refresh_cache-a358a3cb-deda-419a-aa3c-ce7aeb534240" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.086300] env[62600]: DEBUG nova.network.neutron [req-2c3a0c08-47ce-43c4-b146-737338f2e7ac req-509560fa-fb2c-4b03-af5c-083158448b1c service nova] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Refreshing network info cache for port 6a3323e4-5bec-4c2d-8c89-285faea1dd06 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 776.137997] env[62600]: INFO nova.scheduler.client.report [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Deleted allocations for instance 014899cb-db4c-4444-abe1-b0129c0d7db9 [ 776.303413] env[62600]: DEBUG nova.network.neutron [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Successfully created port: 8ca7491d-af98-4a32-b8d9-63e6b57356ef {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 776.383235] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222264, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.467696] env[62600]: DEBUG nova.compute.manager [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 776.523781] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222265, 'name': CreateVM_Task, 'duration_secs': 0.312494} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.526281] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 776.527171] env[62600]: DEBUG oslo_concurrency.lockutils [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.527340] env[62600]: DEBUG oslo_concurrency.lockutils [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.527649] env[62600]: DEBUG oslo_concurrency.lockutils [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 776.528262] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9751e328-bce4-4b70-b420-998c97d8d73b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.532599] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 776.532599] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e9eb63-3c2c-0225-b224-b8e0bdda0c5f" [ 776.532599] env[62600]: _type = "Task" [ 776.532599] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.542093] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e9eb63-3c2c-0225-b224-b8e0bdda0c5f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.552473] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e33e37-638e-940a-df1a-3f047d9d5920, 'name': SearchDatastore_Task, 'duration_secs': 0.009434} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.555419] env[62600]: DEBUG oslo_concurrency.lockutils [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.555697] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] efff4d72-4673-4a8d-9a81-be9d700ff881/efff4d72-4673-4a8d-9a81-be9d700ff881.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 776.556566] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5db73b68-5684-4843-ab7c-138f6e197cdf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.564193] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 776.564193] env[62600]: value = "task-1222266" [ 776.564193] env[62600]: _type = "Task" [ 776.564193] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.575819] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222266, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.648386] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4578a914-94be-4955-9990-3b4c0882583a tempest-ServerActionsV293TestJSON-1805916986 tempest-ServerActionsV293TestJSON-1805916986-project-member] Lock "014899cb-db4c-4444-abe1-b0129c0d7db9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 161.248s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.891277] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dcad09a-7911-42f5-8001-27b953ebdec7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.891899] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222264, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.897610] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d0b97a-4d7d-4b1f-a29e-f050136f102e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.937574] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31942d0a-affd-4a3a-8374-9fb09d26d239 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.946535] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-497f9006-b53c-416e-8df4-072b0dd5d2e6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.963063] env[62600]: DEBUG nova.compute.provider_tree [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.973472] env[62600]: DEBUG nova.network.neutron [req-2c3a0c08-47ce-43c4-b146-737338f2e7ac req-509560fa-fb2c-4b03-af5c-083158448b1c service nova] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Updated VIF entry in instance network info cache for port 6a3323e4-5bec-4c2d-8c89-285faea1dd06. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 776.973472] env[62600]: DEBUG nova.network.neutron [req-2c3a0c08-47ce-43c4-b146-737338f2e7ac req-509560fa-fb2c-4b03-af5c-083158448b1c service nova] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Updating instance_info_cache with network_info: [{"id": "6a3323e4-5bec-4c2d-8c89-285faea1dd06", "address": "fa:16:3e:8e:12:44", "network": {"id": "d946b94e-9827-4c22-af05-072fcaa3d0e6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.176", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ec39f2547f624ed0b240d084b949694a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a3323e4-5b", "ovs_interfaceid": "6a3323e4-5bec-4c2d-8c89-285faea1dd06", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.046281] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e9eb63-3c2c-0225-b224-b8e0bdda0c5f, 'name': SearchDatastore_Task, 'duration_secs': 0.009651} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.046677] env[62600]: DEBUG oslo_concurrency.lockutils [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.047028] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 777.047816] env[62600]: DEBUG oslo_concurrency.lockutils [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.047816] env[62600]: DEBUG oslo_concurrency.lockutils [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.047816] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 777.048160] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-47b2dd85-efa2-47b1-be6a-25ac6d5da516 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.056708] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 777.056941] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 777.057946] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28f61687-b47c-405d-b149-9744cdec1215 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.064451] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 777.064451] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526046ac-4d31-1e9f-37c0-f0e82905b8e0" [ 777.064451] env[62600]: _type = "Task" [ 777.064451] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.080384] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222266, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478246} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.083728] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] efff4d72-4673-4a8d-9a81-be9d700ff881/efff4d72-4673-4a8d-9a81-be9d700ff881.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 777.083728] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 777.083728] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526046ac-4d31-1e9f-37c0-f0e82905b8e0, 'name': SearchDatastore_Task, 'duration_secs': 0.008642} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.083728] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-86c53138-b7df-43d3-b617-23d8c21ca7e8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.086055] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-111818e6-413e-4328-a931-2ffc33b63d8a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.091350] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 777.091350] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521b9e9c-25fa-e19d-304f-4a4f4c0423c4" [ 777.091350] env[62600]: _type = "Task" [ 777.091350] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.092809] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 777.092809] env[62600]: value = "task-1222268" [ 777.092809] env[62600]: _type = "Task" [ 777.092809] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.103977] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521b9e9c-25fa-e19d-304f-4a4f4c0423c4, 'name': SearchDatastore_Task, 'duration_secs': 0.009475} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.106807] env[62600]: DEBUG oslo_concurrency.lockutils [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.107114] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] a358a3cb-deda-419a-aa3c-ce7aeb534240/a358a3cb-deda-419a-aa3c-ce7aeb534240.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 777.107871] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222268, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.108131] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fb287a4a-be46-400a-98f1-57e7e8865f39 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.113952] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 777.113952] env[62600]: value = "task-1222269" [ 777.113952] env[62600]: _type = "Task" [ 777.113952] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.121771] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222269, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.150963] env[62600]: DEBUG nova.compute.manager [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 777.385978] env[62600]: DEBUG oslo_vmware.api [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222264, 'name': PowerOnVM_Task, 'duration_secs': 1.40625} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.386478] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 777.386618] env[62600]: INFO nova.compute.manager [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Took 9.57 seconds to spawn the instance on the hypervisor. [ 777.386798] env[62600]: DEBUG nova.compute.manager [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 777.388257] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a7a1a1-d9a6-4390-b5e0-91857d243e8b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.466765] env[62600]: DEBUG nova.scheduler.client.report [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 777.487560] env[62600]: DEBUG oslo_concurrency.lockutils [req-2c3a0c08-47ce-43c4-b146-737338f2e7ac req-509560fa-fb2c-4b03-af5c-083158448b1c service nova] Releasing lock "refresh_cache-a358a3cb-deda-419a-aa3c-ce7aeb534240" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.489558] env[62600]: DEBUG nova.compute.manager [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 777.532429] env[62600]: DEBUG nova.virt.hardware [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 777.532698] env[62600]: DEBUG nova.virt.hardware [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 777.532883] env[62600]: DEBUG nova.virt.hardware [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 777.533089] env[62600]: DEBUG nova.virt.hardware [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 777.533241] env[62600]: DEBUG nova.virt.hardware [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 777.533385] env[62600]: DEBUG nova.virt.hardware [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 777.533595] env[62600]: DEBUG nova.virt.hardware [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 777.533751] env[62600]: DEBUG nova.virt.hardware [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 777.533918] env[62600]: DEBUG nova.virt.hardware [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 777.534100] env[62600]: DEBUG nova.virt.hardware [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 777.534281] env[62600]: DEBUG nova.virt.hardware [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 777.535196] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deeec08b-7326-4b4a-b5e2-d836121a7ae7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.549614] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89c9ee61-f794-453d-bda0-30ed7be99032 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.607277] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222268, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063003} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.607862] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 777.609028] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c1f3cc1-0202-4b2e-bc8d-07609f8d968d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.642219] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] efff4d72-4673-4a8d-9a81-be9d700ff881/efff4d72-4673-4a8d-9a81-be9d700ff881.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 777.645603] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e385dc51-0246-4ae6-bff5-17eeb822ae88 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.665286] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222269, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.673613] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 777.673613] env[62600]: value = "task-1222270" [ 777.673613] env[62600]: _type = "Task" [ 777.673613] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.683932] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222270, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.685273] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.862199] env[62600]: DEBUG nova.network.neutron [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Successfully updated port: 8ca7491d-af98-4a32-b8d9-63e6b57356ef {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 777.868928] env[62600]: DEBUG nova.compute.manager [req-093f0fc5-00d5-4431-80ab-77c93a1f5292 req-013c13d1-b442-4012-90c1-7d2009311b9c service nova] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Received event network-vif-plugged-8ca7491d-af98-4a32-b8d9-63e6b57356ef {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 777.868928] env[62600]: DEBUG oslo_concurrency.lockutils [req-093f0fc5-00d5-4431-80ab-77c93a1f5292 req-013c13d1-b442-4012-90c1-7d2009311b9c service nova] Acquiring lock "e25c631d-3d1b-40d2-9fb1-a65431f991a9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.868928] env[62600]: DEBUG oslo_concurrency.lockutils [req-093f0fc5-00d5-4431-80ab-77c93a1f5292 req-013c13d1-b442-4012-90c1-7d2009311b9c service nova] Lock "e25c631d-3d1b-40d2-9fb1-a65431f991a9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.868928] env[62600]: DEBUG oslo_concurrency.lockutils [req-093f0fc5-00d5-4431-80ab-77c93a1f5292 req-013c13d1-b442-4012-90c1-7d2009311b9c service nova] Lock "e25c631d-3d1b-40d2-9fb1-a65431f991a9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.868928] env[62600]: DEBUG nova.compute.manager [req-093f0fc5-00d5-4431-80ab-77c93a1f5292 req-013c13d1-b442-4012-90c1-7d2009311b9c service nova] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] No waiting events found dispatching network-vif-plugged-8ca7491d-af98-4a32-b8d9-63e6b57356ef {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 777.869125] env[62600]: WARNING nova.compute.manager [req-093f0fc5-00d5-4431-80ab-77c93a1f5292 req-013c13d1-b442-4012-90c1-7d2009311b9c service nova] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Received unexpected event network-vif-plugged-8ca7491d-af98-4a32-b8d9-63e6b57356ef for instance with vm_state building and task_state spawning. [ 777.908789] env[62600]: INFO nova.compute.manager [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Took 30.42 seconds to build instance. [ 777.978024] env[62600]: DEBUG oslo_concurrency.lockutils [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.514s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.978024] env[62600]: DEBUG nova.compute.manager [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 777.978848] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.909s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.980507] env[62600]: INFO nova.compute.claims [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 778.126237] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222269, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.621159} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.126712] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] a358a3cb-deda-419a-aa3c-ce7aeb534240/a358a3cb-deda-419a-aa3c-ce7aeb534240.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 778.127264] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 778.128088] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d8e1ee98-7a3f-4c4d-972a-59c791946da8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.134064] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 778.134064] env[62600]: value = "task-1222271" [ 778.134064] env[62600]: _type = "Task" [ 778.134064] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.141940] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222271, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.184632] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222270, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.373859] env[62600]: DEBUG oslo_concurrency.lockutils [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Acquiring lock "refresh_cache-e25c631d-3d1b-40d2-9fb1-a65431f991a9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.373859] env[62600]: DEBUG oslo_concurrency.lockutils [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Acquired lock "refresh_cache-e25c631d-3d1b-40d2-9fb1-a65431f991a9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.373859] env[62600]: DEBUG nova.network.neutron [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 778.415041] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85d86bf0-2634-4f06-a252-bdc66cea06aa tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Lock "92cf1acb-b471-4add-a73f-7e9e94fbaaa4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.352s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.480657] env[62600]: DEBUG nova.compute.utils [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 778.482406] env[62600]: DEBUG nova.compute.manager [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 778.482573] env[62600]: DEBUG nova.network.neutron [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 778.539309] env[62600]: DEBUG nova.policy [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7cc09968d9ba41ffb7fdb0a26870f7b0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '880c2a040cca4ee685451a87f8dcab28', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 778.644861] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222271, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.371103} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.645558] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 778.646472] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b35a68e-da6a-4178-9246-fe4c3d492287 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.675598] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Reconfiguring VM instance instance-0000002e to attach disk [datastore1] a358a3cb-deda-419a-aa3c-ce7aeb534240/a358a3cb-deda-419a-aa3c-ce7aeb534240.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 778.676040] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a47f3be0-af98-4046-aaa7-d8f18246be6a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.704126] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222270, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.705523] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 778.705523] env[62600]: value = "task-1222272" [ 778.705523] env[62600]: _type = "Task" [ 778.705523] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.713611] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222272, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.890342] env[62600]: DEBUG nova.network.neutron [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Successfully created port: 194093a1-33d0-42f8-8255-db5802300692 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 778.920409] env[62600]: DEBUG nova.compute.manager [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 778.963514] env[62600]: DEBUG nova.network.neutron [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 778.990447] env[62600]: DEBUG nova.compute.manager [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 779.204051] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222270, 'name': ReconfigVM_Task, 'duration_secs': 1.151161} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.206566] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Reconfigured VM instance instance-0000002d to attach disk [datastore2] efff4d72-4673-4a8d-9a81-be9d700ff881/efff4d72-4673-4a8d-9a81-be9d700ff881.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 779.211018] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bb19e776-3225-4298-8af8-f2639363b39e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.218795] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222272, 'name': ReconfigVM_Task, 'duration_secs': 0.283783} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.220188] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Reconfigured VM instance instance-0000002e to attach disk [datastore1] a358a3cb-deda-419a-aa3c-ce7aeb534240/a358a3cb-deda-419a-aa3c-ce7aeb534240.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 779.221093] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 779.221093] env[62600]: value = "task-1222273" [ 779.221093] env[62600]: _type = "Task" [ 779.221093] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.223190] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-20001142-0bba-497e-94c7-45d443b2abdb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.234180] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222273, 'name': Rename_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.235101] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 779.235101] env[62600]: value = "task-1222274" [ 779.235101] env[62600]: _type = "Task" [ 779.235101] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.246749] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222274, 'name': Rename_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.344877] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248d654d-9e81-45a9-b70e-69bf862d6aee {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.353944] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f9a69e3-3f99-4e62-8ac2-047176420cbe {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.392601] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-610c68b4-c2ae-4c4c-98b7-76ae0493fada {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.401231] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61debfb-a632-4034-a666-e0d422e99030 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.421421] env[62600]: DEBUG nova.compute.provider_tree [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.444197] env[62600]: DEBUG oslo_concurrency.lockutils [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.447399] env[62600]: DEBUG nova.network.neutron [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Updating instance_info_cache with network_info: [{"id": "8ca7491d-af98-4a32-b8d9-63e6b57356ef", "address": "fa:16:3e:1a:e1:b5", "network": {"id": "f27f3669-f56f-4893-ae57-430b121f6309", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-417681146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4e28048248a44a6fb0113b3f84a8e8be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496faa4d-d874-449b-905e-328ddd60b31b", "external-id": "nsx-vlan-transportzone-391", "segmentation_id": 391, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ca7491d-af", "ovs_interfaceid": "8ca7491d-af98-4a32-b8d9-63e6b57356ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.511020] env[62600]: DEBUG nova.compute.manager [req-3a9ea764-5318-483b-b26a-0903729fdcd6 req-1b69ea52-6482-44d6-afb4-d5709f5a953b service nova] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Received event network-changed-779d4e97-daa9-44ec-9ee0-91658a73bc02 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 779.511020] env[62600]: DEBUG nova.compute.manager [req-3a9ea764-5318-483b-b26a-0903729fdcd6 req-1b69ea52-6482-44d6-afb4-d5709f5a953b service nova] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Refreshing instance network info cache due to event network-changed-779d4e97-daa9-44ec-9ee0-91658a73bc02. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 779.511020] env[62600]: DEBUG oslo_concurrency.lockutils [req-3a9ea764-5318-483b-b26a-0903729fdcd6 req-1b69ea52-6482-44d6-afb4-d5709f5a953b service nova] Acquiring lock "refresh_cache-92cf1acb-b471-4add-a73f-7e9e94fbaaa4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.511020] env[62600]: DEBUG oslo_concurrency.lockutils [req-3a9ea764-5318-483b-b26a-0903729fdcd6 req-1b69ea52-6482-44d6-afb4-d5709f5a953b service nova] Acquired lock "refresh_cache-92cf1acb-b471-4add-a73f-7e9e94fbaaa4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.511020] env[62600]: DEBUG nova.network.neutron [req-3a9ea764-5318-483b-b26a-0903729fdcd6 req-1b69ea52-6482-44d6-afb4-d5709f5a953b service nova] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Refreshing network info cache for port 779d4e97-daa9-44ec-9ee0-91658a73bc02 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 779.733801] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222273, 'name': Rename_Task, 'duration_secs': 0.170584} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.733946] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 779.734227] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9b3cc25e-0713-4015-866f-38c9d856619f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.749412] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222274, 'name': Rename_Task, 'duration_secs': 0.143797} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.749412] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 779.749412] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 779.749412] env[62600]: value = "task-1222275" [ 779.749412] env[62600]: _type = "Task" [ 779.749412] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.749693] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c444437c-f93d-4780-a898-bbd96b2fa081 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.759734] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222275, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.761082] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 779.761082] env[62600]: value = "task-1222276" [ 779.761082] env[62600]: _type = "Task" [ 779.761082] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.772348] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222276, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.931120] env[62600]: DEBUG nova.scheduler.client.report [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.953461] env[62600]: DEBUG oslo_concurrency.lockutils [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Releasing lock "refresh_cache-e25c631d-3d1b-40d2-9fb1-a65431f991a9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.953461] env[62600]: DEBUG nova.compute.manager [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Instance network_info: |[{"id": "8ca7491d-af98-4a32-b8d9-63e6b57356ef", "address": "fa:16:3e:1a:e1:b5", "network": {"id": "f27f3669-f56f-4893-ae57-430b121f6309", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-417681146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4e28048248a44a6fb0113b3f84a8e8be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496faa4d-d874-449b-905e-328ddd60b31b", "external-id": "nsx-vlan-transportzone-391", "segmentation_id": 391, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ca7491d-af", "ovs_interfaceid": "8ca7491d-af98-4a32-b8d9-63e6b57356ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 779.953597] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1a:e1:b5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '496faa4d-d874-449b-905e-328ddd60b31b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8ca7491d-af98-4a32-b8d9-63e6b57356ef', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 779.961756] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Creating folder: Project (4e28048248a44a6fb0113b3f84a8e8be). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 779.963314] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ac3dd5ba-156b-4c0b-bef1-b110d06697a4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.976233] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Created folder: Project (4e28048248a44a6fb0113b3f84a8e8be) in parent group-v264198. [ 779.976670] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Creating folder: Instances. Parent ref: group-v264224. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 779.977019] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-04985a2d-0e3f-4f80-8093-56ba75df8f53 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.987508] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Created folder: Instances in parent group-v264224. [ 779.987508] env[62600]: DEBUG oslo.service.loopingcall [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 779.987508] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 779.987508] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7294930c-50b8-4f32-8417-61f5717c795f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.012742] env[62600]: DEBUG nova.compute.manager [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 780.017397] env[62600]: DEBUG nova.compute.manager [req-86728848-6bc9-40f9-8533-fc90439179fb req-f7521222-2df4-473c-ab99-2a93d8bbaa12 service nova] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Received event network-changed-8ca7491d-af98-4a32-b8d9-63e6b57356ef {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 780.017594] env[62600]: DEBUG nova.compute.manager [req-86728848-6bc9-40f9-8533-fc90439179fb req-f7521222-2df4-473c-ab99-2a93d8bbaa12 service nova] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Refreshing instance network info cache due to event network-changed-8ca7491d-af98-4a32-b8d9-63e6b57356ef. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 780.017831] env[62600]: DEBUG oslo_concurrency.lockutils [req-86728848-6bc9-40f9-8533-fc90439179fb req-f7521222-2df4-473c-ab99-2a93d8bbaa12 service nova] Acquiring lock "refresh_cache-e25c631d-3d1b-40d2-9fb1-a65431f991a9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.017990] env[62600]: DEBUG oslo_concurrency.lockutils [req-86728848-6bc9-40f9-8533-fc90439179fb req-f7521222-2df4-473c-ab99-2a93d8bbaa12 service nova] Acquired lock "refresh_cache-e25c631d-3d1b-40d2-9fb1-a65431f991a9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.018206] env[62600]: DEBUG nova.network.neutron [req-86728848-6bc9-40f9-8533-fc90439179fb req-f7521222-2df4-473c-ab99-2a93d8bbaa12 service nova] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Refreshing network info cache for port 8ca7491d-af98-4a32-b8d9-63e6b57356ef {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 780.024549] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 780.024549] env[62600]: value = "task-1222279" [ 780.024549] env[62600]: _type = "Task" [ 780.024549] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.033966] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222279, 'name': CreateVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.048762] env[62600]: DEBUG nova.virt.hardware [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 780.048762] env[62600]: DEBUG nova.virt.hardware [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 780.048762] env[62600]: DEBUG nova.virt.hardware [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 780.049164] env[62600]: DEBUG nova.virt.hardware [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 780.049164] env[62600]: DEBUG nova.virt.hardware [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 780.049164] env[62600]: DEBUG nova.virt.hardware [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 780.049413] env[62600]: DEBUG nova.virt.hardware [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 780.049598] env[62600]: DEBUG nova.virt.hardware [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 780.049815] env[62600]: DEBUG nova.virt.hardware [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 780.050018] env[62600]: DEBUG nova.virt.hardware [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 780.050268] env[62600]: DEBUG nova.virt.hardware [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 780.051639] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a03bb83-c3ca-42fe-8342-fce78b86d8e1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.062403] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f64119-fed7-4f15-aee0-bc6ae3e72fd5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.269653] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222275, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.279961] env[62600]: DEBUG oslo_vmware.api [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222276, 'name': PowerOnVM_Task, 'duration_secs': 0.473761} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.281053] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 780.281053] env[62600]: INFO nova.compute.manager [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Took 7.33 seconds to spawn the instance on the hypervisor. [ 780.281053] env[62600]: DEBUG nova.compute.manager [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 780.282794] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc2cd547-2ea3-4068-bda4-79d231c27c92 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.440033] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.459s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.440033] env[62600]: DEBUG nova.compute.manager [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 780.442024] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.440s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.445652] env[62600]: INFO nova.compute.claims [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 780.457293] env[62600]: DEBUG nova.network.neutron [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Successfully updated port: 194093a1-33d0-42f8-8255-db5802300692 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 780.533624] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222279, 'name': CreateVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.744056] env[62600]: DEBUG nova.network.neutron [req-86728848-6bc9-40f9-8533-fc90439179fb req-f7521222-2df4-473c-ab99-2a93d8bbaa12 service nova] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Updated VIF entry in instance network info cache for port 8ca7491d-af98-4a32-b8d9-63e6b57356ef. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 780.744438] env[62600]: DEBUG nova.network.neutron [req-86728848-6bc9-40f9-8533-fc90439179fb req-f7521222-2df4-473c-ab99-2a93d8bbaa12 service nova] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Updating instance_info_cache with network_info: [{"id": "8ca7491d-af98-4a32-b8d9-63e6b57356ef", "address": "fa:16:3e:1a:e1:b5", "network": {"id": "f27f3669-f56f-4893-ae57-430b121f6309", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-417681146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4e28048248a44a6fb0113b3f84a8e8be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496faa4d-d874-449b-905e-328ddd60b31b", "external-id": "nsx-vlan-transportzone-391", "segmentation_id": 391, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ca7491d-af", "ovs_interfaceid": "8ca7491d-af98-4a32-b8d9-63e6b57356ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.761824] env[62600]: DEBUG oslo_vmware.api [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222275, 'name': PowerOnVM_Task, 'duration_secs': 0.730108} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.762139] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 780.762352] env[62600]: INFO nova.compute.manager [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Took 10.32 seconds to spawn the instance on the hypervisor. [ 780.762532] env[62600]: DEBUG nova.compute.manager [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 780.763468] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e6aa139-ac10-45de-8771-acbaea30de78 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.802630] env[62600]: INFO nova.compute.manager [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Took 26.98 seconds to build instance. [ 780.828636] env[62600]: DEBUG nova.network.neutron [req-3a9ea764-5318-483b-b26a-0903729fdcd6 req-1b69ea52-6482-44d6-afb4-d5709f5a953b service nova] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Updated VIF entry in instance network info cache for port 779d4e97-daa9-44ec-9ee0-91658a73bc02. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 780.829015] env[62600]: DEBUG nova.network.neutron [req-3a9ea764-5318-483b-b26a-0903729fdcd6 req-1b69ea52-6482-44d6-afb4-d5709f5a953b service nova] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Updating instance_info_cache with network_info: [{"id": "779d4e97-daa9-44ec-9ee0-91658a73bc02", "address": "fa:16:3e:d5:8d:e6", "network": {"id": "29b27b28-57ba-466c-a0b5-38469adbe1d5", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1563603335-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49c5a7585f2a4d7888f3240992a68955", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c492f5cc-7ae0-4cab-823c-0d5dd8c60b26", "external-id": "nsx-vlan-transportzone-824", "segmentation_id": 824, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap779d4e97-da", "ovs_interfaceid": "779d4e97-daa9-44ec-9ee0-91658a73bc02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.953232] env[62600]: DEBUG nova.compute.utils [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 780.959499] env[62600]: DEBUG nova.compute.manager [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 780.959499] env[62600]: DEBUG nova.network.neutron [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 780.960582] env[62600]: DEBUG oslo_concurrency.lockutils [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Acquiring lock "refresh_cache-0a8840f1-7681-4aaa-9f3a-32b72c04c1c4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.960668] env[62600]: DEBUG oslo_concurrency.lockutils [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Acquired lock "refresh_cache-0a8840f1-7681-4aaa-9f3a-32b72c04c1c4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.961048] env[62600]: DEBUG nova.network.neutron [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 781.035550] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222279, 'name': CreateVM_Task, 'duration_secs': 0.590125} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.037035] env[62600]: DEBUG nova.policy [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '920b32e4a4b24cd893392587dbd9bf2d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b3573ce4e0a642dc9822f94211bffa69', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 781.038445] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 781.039240] env[62600]: DEBUG oslo_concurrency.lockutils [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.039240] env[62600]: DEBUG oslo_concurrency.lockutils [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.039841] env[62600]: DEBUG oslo_concurrency.lockutils [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 781.040096] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a9e1095-9212-45f7-ad07-bc9eef47cde4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.044519] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Waiting for the task: (returnval){ [ 781.044519] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528410d3-7c43-8115-0f7a-6ce10af7906f" [ 781.044519] env[62600]: _type = "Task" [ 781.044519] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.052818] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528410d3-7c43-8115-0f7a-6ce10af7906f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.247142] env[62600]: DEBUG oslo_concurrency.lockutils [req-86728848-6bc9-40f9-8533-fc90439179fb req-f7521222-2df4-473c-ab99-2a93d8bbaa12 service nova] Releasing lock "refresh_cache-e25c631d-3d1b-40d2-9fb1-a65431f991a9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.283826] env[62600]: INFO nova.compute.manager [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Took 29.40 seconds to build instance. [ 781.302693] env[62600]: DEBUG oslo_concurrency.lockutils [None req-14e8aeec-ba62-43df-9912-1378b65a0658 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lock "a358a3cb-deda-419a-aa3c-ce7aeb534240" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.252s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.331756] env[62600]: DEBUG oslo_concurrency.lockutils [req-3a9ea764-5318-483b-b26a-0903729fdcd6 req-1b69ea52-6482-44d6-afb4-d5709f5a953b service nova] Releasing lock "refresh_cache-92cf1acb-b471-4add-a73f-7e9e94fbaaa4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.457489] env[62600]: DEBUG nova.compute.manager [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 781.495919] env[62600]: DEBUG nova.network.neutron [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 781.501047] env[62600]: DEBUG nova.network.neutron [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Successfully created port: 6c00cb70-1e44-47a1-a1da-6273976cda79 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 781.554664] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528410d3-7c43-8115-0f7a-6ce10af7906f, 'name': SearchDatastore_Task, 'duration_secs': 0.024659} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.555028] env[62600]: DEBUG oslo_concurrency.lockutils [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.555297] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 781.555562] env[62600]: DEBUG oslo_concurrency.lockutils [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.555756] env[62600]: DEBUG oslo_concurrency.lockutils [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.555983] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 781.556279] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b6169741-c527-413c-ae83-b0aae5696c7c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.568185] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 781.568511] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 781.569200] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b37ce2b-f198-486c-8016-e8da99c49591 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.574374] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Waiting for the task: (returnval){ [ 781.574374] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d86926-e334-5300-2a05-ae6199a68f5b" [ 781.574374] env[62600]: _type = "Task" [ 781.574374] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.585121] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d86926-e334-5300-2a05-ae6199a68f5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.645403] env[62600]: DEBUG nova.network.neutron [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Updating instance_info_cache with network_info: [{"id": "194093a1-33d0-42f8-8255-db5802300692", "address": "fa:16:3e:4e:df:59", "network": {"id": "884e4182-92a8-4da3-80bc-b98a3fb9b0ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-883977721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "880c2a040cca4ee685451a87f8dcab28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap194093a1-33", "ovs_interfaceid": "194093a1-33d0-42f8-8255-db5802300692", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.785850] env[62600]: DEBUG oslo_concurrency.lockutils [None req-96acb54b-ce2c-42c4-a7a7-c48ddb79959c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lock "efff4d72-4673-4a8d-9a81-be9d700ff881" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.404s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.806718] env[62600]: DEBUG nova.compute.manager [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 781.838298] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6433272-0a5d-47db-aefe-802933eef5b3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.847125] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63025f86-1a68-4725-8a70-b955bf77dcce {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.883626] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f6e4105-f2c6-4459-8887-ac02e7534005 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.891533] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df027f5-5278-44f3-8123-e77f4503d449 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.906532] env[62600]: DEBUG nova.compute.provider_tree [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.088691] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d86926-e334-5300-2a05-ae6199a68f5b, 'name': SearchDatastore_Task, 'duration_secs': 0.013439} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.089578] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38c21409-1f18-4433-81ce-5b079d13f0b3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.094854] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Waiting for the task: (returnval){ [ 782.094854] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52df2325-73b1-51b7-0ad0-82a8bceeec3d" [ 782.094854] env[62600]: _type = "Task" [ 782.094854] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.103920] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52df2325-73b1-51b7-0ad0-82a8bceeec3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.151164] env[62600]: DEBUG oslo_concurrency.lockutils [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Releasing lock "refresh_cache-0a8840f1-7681-4aaa-9f3a-32b72c04c1c4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.151488] env[62600]: DEBUG nova.compute.manager [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Instance network_info: |[{"id": "194093a1-33d0-42f8-8255-db5802300692", "address": "fa:16:3e:4e:df:59", "network": {"id": "884e4182-92a8-4da3-80bc-b98a3fb9b0ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-883977721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "880c2a040cca4ee685451a87f8dcab28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap194093a1-33", "ovs_interfaceid": "194093a1-33d0-42f8-8255-db5802300692", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 782.151889] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4e:df:59', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e1c9bb98-73a9-48eb-856e-a541afe9b07b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '194093a1-33d0-42f8-8255-db5802300692', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 782.159988] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Creating folder: Project (880c2a040cca4ee685451a87f8dcab28). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 782.160288] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cfbea119-1180-4664-9846-ba1f5238e61d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.168785] env[62600]: DEBUG nova.compute.manager [req-bb5c225a-ef45-4535-aabd-5300e6d783c0 req-33f09287-1146-4a91-8257-53151ae27a97 service nova] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Received event network-vif-plugged-194093a1-33d0-42f8-8255-db5802300692 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 782.169122] env[62600]: DEBUG oslo_concurrency.lockutils [req-bb5c225a-ef45-4535-aabd-5300e6d783c0 req-33f09287-1146-4a91-8257-53151ae27a97 service nova] Acquiring lock "0a8840f1-7681-4aaa-9f3a-32b72c04c1c4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.169207] env[62600]: DEBUG oslo_concurrency.lockutils [req-bb5c225a-ef45-4535-aabd-5300e6d783c0 req-33f09287-1146-4a91-8257-53151ae27a97 service nova] Lock "0a8840f1-7681-4aaa-9f3a-32b72c04c1c4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.169381] env[62600]: DEBUG oslo_concurrency.lockutils [req-bb5c225a-ef45-4535-aabd-5300e6d783c0 req-33f09287-1146-4a91-8257-53151ae27a97 service nova] Lock "0a8840f1-7681-4aaa-9f3a-32b72c04c1c4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.169551] env[62600]: DEBUG nova.compute.manager [req-bb5c225a-ef45-4535-aabd-5300e6d783c0 req-33f09287-1146-4a91-8257-53151ae27a97 service nova] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] No waiting events found dispatching network-vif-plugged-194093a1-33d0-42f8-8255-db5802300692 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 782.169711] env[62600]: WARNING nova.compute.manager [req-bb5c225a-ef45-4535-aabd-5300e6d783c0 req-33f09287-1146-4a91-8257-53151ae27a97 service nova] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Received unexpected event network-vif-plugged-194093a1-33d0-42f8-8255-db5802300692 for instance with vm_state building and task_state spawning. [ 782.169863] env[62600]: DEBUG nova.compute.manager [req-bb5c225a-ef45-4535-aabd-5300e6d783c0 req-33f09287-1146-4a91-8257-53151ae27a97 service nova] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Received event network-changed-194093a1-33d0-42f8-8255-db5802300692 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 782.170025] env[62600]: DEBUG nova.compute.manager [req-bb5c225a-ef45-4535-aabd-5300e6d783c0 req-33f09287-1146-4a91-8257-53151ae27a97 service nova] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Refreshing instance network info cache due to event network-changed-194093a1-33d0-42f8-8255-db5802300692. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 782.170210] env[62600]: DEBUG oslo_concurrency.lockutils [req-bb5c225a-ef45-4535-aabd-5300e6d783c0 req-33f09287-1146-4a91-8257-53151ae27a97 service nova] Acquiring lock "refresh_cache-0a8840f1-7681-4aaa-9f3a-32b72c04c1c4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.170457] env[62600]: DEBUG oslo_concurrency.lockutils [req-bb5c225a-ef45-4535-aabd-5300e6d783c0 req-33f09287-1146-4a91-8257-53151ae27a97 service nova] Acquired lock "refresh_cache-0a8840f1-7681-4aaa-9f3a-32b72c04c1c4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.171444] env[62600]: DEBUG nova.network.neutron [req-bb5c225a-ef45-4535-aabd-5300e6d783c0 req-33f09287-1146-4a91-8257-53151ae27a97 service nova] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Refreshing network info cache for port 194093a1-33d0-42f8-8255-db5802300692 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 782.173028] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Created folder: Project (880c2a040cca4ee685451a87f8dcab28) in parent group-v264198. [ 782.173155] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Creating folder: Instances. Parent ref: group-v264227. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 782.173804] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6c834444-6d6f-4dc7-b316-24274db2f1b3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.182881] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Created folder: Instances in parent group-v264227. [ 782.183120] env[62600]: DEBUG oslo.service.loopingcall [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.183302] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 782.183712] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-94ccae49-35d0-46e3-bc8b-f3a52650d644 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.205072] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 782.205072] env[62600]: value = "task-1222282" [ 782.205072] env[62600]: _type = "Task" [ 782.205072] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.212577] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222282, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.289048] env[62600]: DEBUG nova.compute.manager [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 782.332987] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.410998] env[62600]: DEBUG nova.scheduler.client.report [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 782.477028] env[62600]: DEBUG nova.compute.manager [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 782.502731] env[62600]: DEBUG nova.virt.hardware [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 782.502968] env[62600]: DEBUG nova.virt.hardware [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 782.503135] env[62600]: DEBUG nova.virt.hardware [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 782.503322] env[62600]: DEBUG nova.virt.hardware [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 782.503465] env[62600]: DEBUG nova.virt.hardware [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 782.503608] env[62600]: DEBUG nova.virt.hardware [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 782.503822] env[62600]: DEBUG nova.virt.hardware [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 782.503988] env[62600]: DEBUG nova.virt.hardware [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 782.504167] env[62600]: DEBUG nova.virt.hardware [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 782.504330] env[62600]: DEBUG nova.virt.hardware [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 782.504509] env[62600]: DEBUG nova.virt.hardware [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 782.505394] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f80d5fe9-1374-4b9d-8dd2-dd9a794a5b89 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.515456] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-204c50e2-31ab-442c-8b81-e6d5b55a1596 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.607480] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52df2325-73b1-51b7-0ad0-82a8bceeec3d, 'name': SearchDatastore_Task, 'duration_secs': 0.012016} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.607754] env[62600]: DEBUG oslo_concurrency.lockutils [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.608022] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] e25c631d-3d1b-40d2-9fb1-a65431f991a9/e25c631d-3d1b-40d2-9fb1-a65431f991a9.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 782.608345] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-158e81fd-1174-4d50-b5fa-c120e9b53105 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.615633] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Waiting for the task: (returnval){ [ 782.615633] env[62600]: value = "task-1222283" [ 782.615633] env[62600]: _type = "Task" [ 782.615633] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.624904] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': task-1222283, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.715042] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222282, 'name': CreateVM_Task, 'duration_secs': 0.316344} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.715042] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 782.715679] env[62600]: DEBUG oslo_concurrency.lockutils [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.715850] env[62600]: DEBUG oslo_concurrency.lockutils [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.716300] env[62600]: DEBUG oslo_concurrency.lockutils [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 782.718742] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6b7e3ff-f3b0-4444-8c43-30012cb9a0d2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.723452] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Waiting for the task: (returnval){ [ 782.723452] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5264b82e-8121-aa0e-dc40-083229bec4f5" [ 782.723452] env[62600]: _type = "Task" [ 782.723452] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.733131] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5264b82e-8121-aa0e-dc40-083229bec4f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.813125] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.920820] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.921466] env[62600]: DEBUG nova.compute.manager [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 782.925227] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 14.404s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.925472] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.925655] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62600) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 782.925986] env[62600]: DEBUG oslo_concurrency.lockutils [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.090s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.927849] env[62600]: INFO nova.compute.claims [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 782.931597] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eea984d4-a9da-4f5a-82a7-b364c68910ad {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.943672] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba4c0db-30e4-44b1-8687-9a56a7e599a2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.960870] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d2a7451-9bb2-4152-89a9-d4dca47ac641 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.971168] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c9ff31-4f57-4636-8788-eaf63603f5ca {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.014991] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181513MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=62600) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 783.015461] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.057037] env[62600]: DEBUG nova.network.neutron [req-bb5c225a-ef45-4535-aabd-5300e6d783c0 req-33f09287-1146-4a91-8257-53151ae27a97 service nova] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Updated VIF entry in instance network info cache for port 194093a1-33d0-42f8-8255-db5802300692. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 783.057037] env[62600]: DEBUG nova.network.neutron [req-bb5c225a-ef45-4535-aabd-5300e6d783c0 req-33f09287-1146-4a91-8257-53151ae27a97 service nova] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Updating instance_info_cache with network_info: [{"id": "194093a1-33d0-42f8-8255-db5802300692", "address": "fa:16:3e:4e:df:59", "network": {"id": "884e4182-92a8-4da3-80bc-b98a3fb9b0ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-883977721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "880c2a040cca4ee685451a87f8dcab28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap194093a1-33", "ovs_interfaceid": "194093a1-33d0-42f8-8255-db5802300692", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.127329] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': task-1222283, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485238} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.127329] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] e25c631d-3d1b-40d2-9fb1-a65431f991a9/e25c631d-3d1b-40d2-9fb1-a65431f991a9.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 783.127329] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 783.127329] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-528e7ccc-38ec-45ee-86ca-03fb4108b79b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.132430] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Waiting for the task: (returnval){ [ 783.132430] env[62600]: value = "task-1222284" [ 783.132430] env[62600]: _type = "Task" [ 783.132430] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.140874] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': task-1222284, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.236020] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5264b82e-8121-aa0e-dc40-083229bec4f5, 'name': SearchDatastore_Task, 'duration_secs': 0.0091} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.236478] env[62600]: DEBUG oslo_concurrency.lockutils [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.236809] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.237153] env[62600]: DEBUG oslo_concurrency.lockutils [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.237451] env[62600]: DEBUG oslo_concurrency.lockutils [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.237586] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.237894] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c38b0700-551a-4afe-8e0f-e0f0274637e6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.248466] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.248728] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 783.249626] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69ee4f06-6442-4420-b43d-64284a80f751 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.255392] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Waiting for the task: (returnval){ [ 783.255392] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]529770b3-e440-a667-64ed-4e158fa83991" [ 783.255392] env[62600]: _type = "Task" [ 783.255392] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.264736] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]529770b3-e440-a667-64ed-4e158fa83991, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.384958] env[62600]: DEBUG nova.compute.manager [req-b866441e-1be9-4095-8a12-258ea93afa72 req-c57a3eff-eaa9-4ffd-82e3-694a170ca73e service nova] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Received event network-vif-plugged-6c00cb70-1e44-47a1-a1da-6273976cda79 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 783.385358] env[62600]: DEBUG oslo_concurrency.lockutils [req-b866441e-1be9-4095-8a12-258ea93afa72 req-c57a3eff-eaa9-4ffd-82e3-694a170ca73e service nova] Acquiring lock "dcbe8e2d-6ae9-465f-8394-3978ee61a15b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.385454] env[62600]: DEBUG oslo_concurrency.lockutils [req-b866441e-1be9-4095-8a12-258ea93afa72 req-c57a3eff-eaa9-4ffd-82e3-694a170ca73e service nova] Lock "dcbe8e2d-6ae9-465f-8394-3978ee61a15b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.385942] env[62600]: DEBUG oslo_concurrency.lockutils [req-b866441e-1be9-4095-8a12-258ea93afa72 req-c57a3eff-eaa9-4ffd-82e3-694a170ca73e service nova] Lock "dcbe8e2d-6ae9-465f-8394-3978ee61a15b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.385942] env[62600]: DEBUG nova.compute.manager [req-b866441e-1be9-4095-8a12-258ea93afa72 req-c57a3eff-eaa9-4ffd-82e3-694a170ca73e service nova] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] No waiting events found dispatching network-vif-plugged-6c00cb70-1e44-47a1-a1da-6273976cda79 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 783.385942] env[62600]: WARNING nova.compute.manager [req-b866441e-1be9-4095-8a12-258ea93afa72 req-c57a3eff-eaa9-4ffd-82e3-694a170ca73e service nova] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Received unexpected event network-vif-plugged-6c00cb70-1e44-47a1-a1da-6273976cda79 for instance with vm_state building and task_state spawning. [ 783.432680] env[62600]: DEBUG nova.compute.utils [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 783.434493] env[62600]: DEBUG nova.compute.manager [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Not allocating networking since 'none' was specified. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 783.441768] env[62600]: DEBUG nova.network.neutron [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Successfully updated port: 6c00cb70-1e44-47a1-a1da-6273976cda79 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 783.531723] env[62600]: DEBUG nova.compute.manager [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 783.533686] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cd2ff34-56d1-4d67-994f-7523ef1c1020 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.559140] env[62600]: DEBUG oslo_concurrency.lockutils [req-bb5c225a-ef45-4535-aabd-5300e6d783c0 req-33f09287-1146-4a91-8257-53151ae27a97 service nova] Releasing lock "refresh_cache-0a8840f1-7681-4aaa-9f3a-32b72c04c1c4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.645468] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': task-1222284, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069473} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.645928] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 783.649319] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27538cc-6db4-493d-927a-783ae09cb6f8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.674037] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] e25c631d-3d1b-40d2-9fb1-a65431f991a9/e25c631d-3d1b-40d2-9fb1-a65431f991a9.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 783.674334] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3c8955d-d909-4e98-8912-993185ae50e7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.696172] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Waiting for the task: (returnval){ [ 783.696172] env[62600]: value = "task-1222285" [ 783.696172] env[62600]: _type = "Task" [ 783.696172] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.704016] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': task-1222285, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.766119] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]529770b3-e440-a667-64ed-4e158fa83991, 'name': SearchDatastore_Task, 'duration_secs': 0.013589} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.766926] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26a1c662-6a06-4b0d-9fe0-35119bc7a0dc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.771741] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Waiting for the task: (returnval){ [ 783.771741] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c8097b-2124-8333-64c5-71a94363e68c" [ 783.771741] env[62600]: _type = "Task" [ 783.771741] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.779530] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c8097b-2124-8333-64c5-71a94363e68c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.937177] env[62600]: DEBUG nova.compute.manager [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 783.950267] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Acquiring lock "refresh_cache-dcbe8e2d-6ae9-465f-8394-3978ee61a15b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.950267] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Acquired lock "refresh_cache-dcbe8e2d-6ae9-465f-8394-3978ee61a15b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.950267] env[62600]: DEBUG nova.network.neutron [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 784.050052] env[62600]: INFO nova.compute.manager [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] instance snapshotting [ 784.055510] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f30dc8-a5cc-4ef6-bcbc-d6a18b1609b6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.079944] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-107afee6-56d8-4559-834a-0c98be5390e7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.208353] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': task-1222285, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.284156] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c8097b-2124-8333-64c5-71a94363e68c, 'name': SearchDatastore_Task, 'duration_secs': 0.090976} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.284736] env[62600]: DEBUG oslo_concurrency.lockutils [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.285036] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4/0a8840f1-7681-4aaa-9f3a-32b72c04c1c4.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 784.285300] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a2d5a1e8-7ecc-4e24-9c47-e4f130fc239e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.294623] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Waiting for the task: (returnval){ [ 784.294623] env[62600]: value = "task-1222286" [ 784.294623] env[62600]: _type = "Task" [ 784.294623] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.304013] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': task-1222286, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.351019] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31bf0806-e27c-462e-84ad-db37564ec8ba {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.357497] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d4d6adf-9544-4bc4-a7c7-e3121c73a20a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.393199] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb6a0fc0-2863-4088-b049-41b81f0ee865 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.400342] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13ee67d-ca87-429d-92c3-249202451c50 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.414650] env[62600]: DEBUG nova.compute.provider_tree [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.506677] env[62600]: DEBUG nova.network.neutron [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 784.590900] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Creating Snapshot of the VM instance {{(pid=62600) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 784.591232] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2d4f2044-54a8-4304-978b-814931176842 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.598800] env[62600]: DEBUG oslo_vmware.api [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 784.598800] env[62600]: value = "task-1222287" [ 784.598800] env[62600]: _type = "Task" [ 784.598800] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.614507] env[62600]: DEBUG oslo_vmware.api [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222287, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.712271] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': task-1222285, 'name': ReconfigVM_Task, 'duration_secs': 0.980255} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.713195] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Reconfigured VM instance instance-0000002f to attach disk [datastore1] e25c631d-3d1b-40d2-9fb1-a65431f991a9/e25c631d-3d1b-40d2-9fb1-a65431f991a9.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 784.714026] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f1c45bfc-5f8b-49f3-939b-26862fb4a314 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.722428] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Waiting for the task: (returnval){ [ 784.722428] env[62600]: value = "task-1222288" [ 784.722428] env[62600]: _type = "Task" [ 784.722428] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.735160] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': task-1222288, 'name': Rename_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.799300] env[62600]: DEBUG nova.network.neutron [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Updating instance_info_cache with network_info: [{"id": "6c00cb70-1e44-47a1-a1da-6273976cda79", "address": "fa:16:3e:bd:c8:f2", "network": {"id": "21b9f8ee-813d-4a89-8b73-39fcb609e331", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1518108585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3573ce4e0a642dc9822f94211bffa69", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c00cb70-1e", "ovs_interfaceid": "6c00cb70-1e44-47a1-a1da-6273976cda79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.811652] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': task-1222286, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.921025] env[62600]: DEBUG nova.scheduler.client.report [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 784.951410] env[62600]: DEBUG nova.compute.manager [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 784.985115] env[62600]: DEBUG nova.virt.hardware [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 784.985684] env[62600]: DEBUG nova.virt.hardware [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 784.985684] env[62600]: DEBUG nova.virt.hardware [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 784.986425] env[62600]: DEBUG nova.virt.hardware [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 784.986425] env[62600]: DEBUG nova.virt.hardware [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 784.986603] env[62600]: DEBUG nova.virt.hardware [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 784.986821] env[62600]: DEBUG nova.virt.hardware [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 784.987110] env[62600]: DEBUG nova.virt.hardware [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 784.987363] env[62600]: DEBUG nova.virt.hardware [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 784.987622] env[62600]: DEBUG nova.virt.hardware [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 784.987860] env[62600]: DEBUG nova.virt.hardware [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 784.989640] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5487f6eb-bc04-458d-9439-7a9f710eb05b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.998768] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d9a4f6-a205-438f-90c6-5e28dd5c868e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.016775] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Instance VIF info [] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 785.023160] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Creating folder: Project (48cae62f463a47f08c6ea0a5403f41c3). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 785.023967] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ba644531-05f3-4eda-ae26-af869ca79301 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.033996] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Created folder: Project (48cae62f463a47f08c6ea0a5403f41c3) in parent group-v264198. [ 785.034251] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Creating folder: Instances. Parent ref: group-v264230. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 785.034591] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-00592bb2-a5c3-4f4c-baa5-f27f1e592882 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.044442] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Created folder: Instances in parent group-v264230. [ 785.044698] env[62600]: DEBUG oslo.service.loopingcall [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.044888] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 785.045100] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f5578d76-ba80-4a3c-bdec-69c596a22b84 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.061768] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 785.061768] env[62600]: value = "task-1222291" [ 785.061768] env[62600]: _type = "Task" [ 785.061768] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.070490] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222291, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.108713] env[62600]: DEBUG oslo_vmware.api [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222287, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.233311] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': task-1222288, 'name': Rename_Task, 'duration_secs': 0.23188} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.233578] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 785.233808] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-19fa4737-bf06-4499-8361-9896c6dfcebd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.241088] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Waiting for the task: (returnval){ [ 785.241088] env[62600]: value = "task-1222292" [ 785.241088] env[62600]: _type = "Task" [ 785.241088] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.248349] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': task-1222292, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.305517] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': task-1222286, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.784242} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.305790] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4/0a8840f1-7681-4aaa-9f3a-32b72c04c1c4.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 785.306157] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 785.306278] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7712d9f7-39e4-4da9-82e6-c2b1f9f29b88 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.308456] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Releasing lock "refresh_cache-dcbe8e2d-6ae9-465f-8394-3978ee61a15b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.310928] env[62600]: DEBUG nova.compute.manager [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Instance network_info: |[{"id": "6c00cb70-1e44-47a1-a1da-6273976cda79", "address": "fa:16:3e:bd:c8:f2", "network": {"id": "21b9f8ee-813d-4a89-8b73-39fcb609e331", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1518108585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3573ce4e0a642dc9822f94211bffa69", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c00cb70-1e", "ovs_interfaceid": "6c00cb70-1e44-47a1-a1da-6273976cda79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 785.311074] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bd:c8:f2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6c00cb70-1e44-47a1-a1da-6273976cda79', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 785.316926] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Creating folder: Project (b3573ce4e0a642dc9822f94211bffa69). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 785.317776] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7b826ed8-9520-4692-a809-07c78e6972d1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.321887] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Waiting for the task: (returnval){ [ 785.321887] env[62600]: value = "task-1222293" [ 785.321887] env[62600]: _type = "Task" [ 785.321887] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.330021] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': task-1222293, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.331637] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Created folder: Project (b3573ce4e0a642dc9822f94211bffa69) in parent group-v264198. [ 785.331928] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Creating folder: Instances. Parent ref: group-v264233. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 785.332310] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53042178-3586-4def-800d-59edd6ed5fee {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.341653] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Created folder: Instances in parent group-v264233. [ 785.341727] env[62600]: DEBUG oslo.service.loopingcall [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.341912] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 785.342138] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-412eae8c-7648-4cda-9ae8-67b58b5ac955 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.361346] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 785.361346] env[62600]: value = "task-1222296" [ 785.361346] env[62600]: _type = "Task" [ 785.361346] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.371943] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222296, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.426592] env[62600]: DEBUG nova.compute.manager [req-f65e6f3c-6661-40b6-a2f4-d29cca222727 req-100a90ac-a096-4c27-8a31-2751aeb539cb service nova] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Received event network-changed-6c00cb70-1e44-47a1-a1da-6273976cda79 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 785.426797] env[62600]: DEBUG nova.compute.manager [req-f65e6f3c-6661-40b6-a2f4-d29cca222727 req-100a90ac-a096-4c27-8a31-2751aeb539cb service nova] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Refreshing instance network info cache due to event network-changed-6c00cb70-1e44-47a1-a1da-6273976cda79. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 785.427065] env[62600]: DEBUG oslo_concurrency.lockutils [req-f65e6f3c-6661-40b6-a2f4-d29cca222727 req-100a90ac-a096-4c27-8a31-2751aeb539cb service nova] Acquiring lock "refresh_cache-dcbe8e2d-6ae9-465f-8394-3978ee61a15b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.427742] env[62600]: DEBUG oslo_concurrency.lockutils [req-f65e6f3c-6661-40b6-a2f4-d29cca222727 req-100a90ac-a096-4c27-8a31-2751aeb539cb service nova] Acquired lock "refresh_cache-dcbe8e2d-6ae9-465f-8394-3978ee61a15b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.427742] env[62600]: DEBUG nova.network.neutron [req-f65e6f3c-6661-40b6-a2f4-d29cca222727 req-100a90ac-a096-4c27-8a31-2751aeb539cb service nova] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Refreshing network info cache for port 6c00cb70-1e44-47a1-a1da-6273976cda79 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 785.429152] env[62600]: DEBUG oslo_concurrency.lockutils [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.503s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.429596] env[62600]: DEBUG nova.compute.manager [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 785.433818] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.193s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.434621] env[62600]: INFO nova.compute.claims [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 785.535597] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Acquiring lock "ddc80377-a14e-48d0-ad24-cc6daf7f6850" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.536035] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Lock "ddc80377-a14e-48d0-ad24-cc6daf7f6850" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.573773] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222291, 'name': CreateVM_Task} progress is 25%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.611777] env[62600]: DEBUG oslo_vmware.api [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222287, 'name': CreateSnapshot_Task, 'duration_secs': 0.986535} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.612098] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Created Snapshot of the VM instance {{(pid=62600) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 785.612848] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4006a28d-b164-45a3-8f57-112efceec683 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.750696] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': task-1222292, 'name': PowerOnVM_Task} progress is 81%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.831150] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': task-1222293, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065165} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.831646] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 785.832412] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39e8ad19-e133-45b8-b6fb-8917a4700b27 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.856056] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4/0a8840f1-7681-4aaa-9f3a-32b72c04c1c4.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 785.856389] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c91aa969-81af-48bd-84b2-c9e40397b5e8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.886811] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222296, 'name': CreateVM_Task} progress is 25%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.888267] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Waiting for the task: (returnval){ [ 785.888267] env[62600]: value = "task-1222297" [ 785.888267] env[62600]: _type = "Task" [ 785.888267] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.895970] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': task-1222297, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.939435] env[62600]: DEBUG nova.compute.utils [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 785.942590] env[62600]: DEBUG nova.compute.manager [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Not allocating networking since 'none' was specified. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 786.077147] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222291, 'name': CreateVM_Task, 'duration_secs': 0.593083} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.077413] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 786.077913] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.078149] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.079265] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 786.080270] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fc22f6c-98f4-49e3-97c4-96bc8097978a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.085663] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 786.085663] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d4f7a1-326e-39ef-0a6b-72ce7a20cf80" [ 786.085663] env[62600]: _type = "Task" [ 786.085663] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.095328] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d4f7a1-326e-39ef-0a6b-72ce7a20cf80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.134551] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Creating linked-clone VM from snapshot {{(pid=62600) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 786.136975] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-340adc46-667b-40c5-b14a-238305c01efc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.144802] env[62600]: DEBUG oslo_vmware.api [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 786.144802] env[62600]: value = "task-1222298" [ 786.144802] env[62600]: _type = "Task" [ 786.144802] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.155261] env[62600]: DEBUG oslo_vmware.api [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222298, 'name': CloneVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.252717] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': task-1222292, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.268042] env[62600]: DEBUG nova.network.neutron [req-f65e6f3c-6661-40b6-a2f4-d29cca222727 req-100a90ac-a096-4c27-8a31-2751aeb539cb service nova] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Updated VIF entry in instance network info cache for port 6c00cb70-1e44-47a1-a1da-6273976cda79. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 786.268492] env[62600]: DEBUG nova.network.neutron [req-f65e6f3c-6661-40b6-a2f4-d29cca222727 req-100a90ac-a096-4c27-8a31-2751aeb539cb service nova] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Updating instance_info_cache with network_info: [{"id": "6c00cb70-1e44-47a1-a1da-6273976cda79", "address": "fa:16:3e:bd:c8:f2", "network": {"id": "21b9f8ee-813d-4a89-8b73-39fcb609e331", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1518108585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b3573ce4e0a642dc9822f94211bffa69", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6c00cb70-1e", "ovs_interfaceid": "6c00cb70-1e44-47a1-a1da-6273976cda79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.389931] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222296, 'name': CreateVM_Task, 'duration_secs': 0.745431} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.394536] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 786.395277] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.400226] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': task-1222297, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.443753] env[62600]: DEBUG nova.compute.manager [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 786.595551] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d4f7a1-326e-39ef-0a6b-72ce7a20cf80, 'name': SearchDatastore_Task, 'duration_secs': 0.026397} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.598300] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.598609] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 786.598851] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.599010] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.599196] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 786.599737] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.600042] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 786.600281] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-623a9ca7-775f-4df5-8a9d-75d74e50103a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.602240] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c109a49-bcf0-4c1c-b3f9-f972aa3d0998 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.607043] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Waiting for the task: (returnval){ [ 786.607043] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]524801d8-8634-9f92-b3f6-a8545f20244b" [ 786.607043] env[62600]: _type = "Task" [ 786.607043] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.613682] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 786.613682] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 786.615259] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26ba2b62-4057-4575-bff8-dee9898e31d0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.620439] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]524801d8-8634-9f92-b3f6-a8545f20244b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.623238] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 786.623238] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52516878-4fda-4a92-33dd-906e9ec55e84" [ 786.623238] env[62600]: _type = "Task" [ 786.623238] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.633324] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52516878-4fda-4a92-33dd-906e9ec55e84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.655771] env[62600]: DEBUG oslo_vmware.api [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222298, 'name': CloneVM_Task} progress is 93%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.754034] env[62600]: DEBUG oslo_vmware.api [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': task-1222292, 'name': PowerOnVM_Task, 'duration_secs': 1.041361} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.754369] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 786.754536] env[62600]: INFO nova.compute.manager [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Took 9.27 seconds to spawn the instance on the hypervisor. [ 786.754726] env[62600]: DEBUG nova.compute.manager [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 786.755576] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8fdf81d-5ee4-4551-9307-8cb2db2b07d3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.771370] env[62600]: DEBUG oslo_concurrency.lockutils [req-f65e6f3c-6661-40b6-a2f4-d29cca222727 req-100a90ac-a096-4c27-8a31-2751aeb539cb service nova] Releasing lock "refresh_cache-dcbe8e2d-6ae9-465f-8394-3978ee61a15b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.855727] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2de040d-8be9-4114-aeb1-7a00b026b5ce {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.864178] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3668f769-c017-4e30-abe8-11dfc4d0d622 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.902594] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8742743-50a2-4b02-889b-f1d6066df49b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.913077] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3721031-7f28-4b56-a6c0-b59a876ad4f0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.916931] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': task-1222297, 'name': ReconfigVM_Task, 'duration_secs': 0.569044} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.917773] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Reconfigured VM instance instance-00000030 to attach disk [datastore2] 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4/0a8840f1-7681-4aaa-9f3a-32b72c04c1c4.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 786.918813] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0b9e972b-b865-407b-a49e-2e5c683201ae {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.930588] env[62600]: DEBUG nova.compute.provider_tree [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.933754] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Waiting for the task: (returnval){ [ 786.933754] env[62600]: value = "task-1222299" [ 786.933754] env[62600]: _type = "Task" [ 786.933754] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.942562] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': task-1222299, 'name': Rename_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.118026] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]524801d8-8634-9f92-b3f6-a8545f20244b, 'name': SearchDatastore_Task, 'duration_secs': 0.012413} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.118153] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.118394] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 787.118624] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.133741] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52516878-4fda-4a92-33dd-906e9ec55e84, 'name': SearchDatastore_Task, 'duration_secs': 0.00954} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.135030] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b1e5660-a328-4016-bb94-97626e0447d9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.140427] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 787.140427] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]529b6a4f-9155-64eb-1842-68258455c122" [ 787.140427] env[62600]: _type = "Task" [ 787.140427] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.147731] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]529b6a4f-9155-64eb-1842-68258455c122, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.154942] env[62600]: DEBUG oslo_vmware.api [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222298, 'name': CloneVM_Task} progress is 94%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.277961] env[62600]: INFO nova.compute.manager [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Took 29.03 seconds to build instance. [ 787.435908] env[62600]: DEBUG nova.scheduler.client.report [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 787.448901] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': task-1222299, 'name': Rename_Task, 'duration_secs': 0.149231} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.449226] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 787.449471] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-57f19a02-73b1-4141-bf85-ea08dc408810 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.453584] env[62600]: DEBUG nova.compute.manager [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 787.456871] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Waiting for the task: (returnval){ [ 787.456871] env[62600]: value = "task-1222300" [ 787.456871] env[62600]: _type = "Task" [ 787.456871] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.471177] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': task-1222300, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.476906] env[62600]: DEBUG nova.virt.hardware [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 787.477175] env[62600]: DEBUG nova.virt.hardware [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 787.477346] env[62600]: DEBUG nova.virt.hardware [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 787.477531] env[62600]: DEBUG nova.virt.hardware [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 787.477677] env[62600]: DEBUG nova.virt.hardware [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 787.477819] env[62600]: DEBUG nova.virt.hardware [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 787.478035] env[62600]: DEBUG nova.virt.hardware [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 787.478199] env[62600]: DEBUG nova.virt.hardware [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 787.478374] env[62600]: DEBUG nova.virt.hardware [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 787.478601] env[62600]: DEBUG nova.virt.hardware [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 787.478806] env[62600]: DEBUG nova.virt.hardware [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 787.479626] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-107c86e3-7e5e-4120-aab2-41bc248911f3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.486844] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82516917-cb68-42e9-8c00-8b28e2ac3ec2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.501617] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Instance VIF info [] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 787.507559] env[62600]: DEBUG oslo.service.loopingcall [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.507818] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 787.508054] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e8fe7df0-6b01-4bdf-9454-ab9ae57a84ff {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.525152] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 787.525152] env[62600]: value = "task-1222301" [ 787.525152] env[62600]: _type = "Task" [ 787.525152] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.534968] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222301, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.655393] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]529b6a4f-9155-64eb-1842-68258455c122, 'name': SearchDatastore_Task, 'duration_secs': 0.009002} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.657034] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.657034] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc/b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 787.657034] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.657034] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 787.657314] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e86c1b6a-40c3-41a4-b535-cdac04bd43fe {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.662381] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e99ef100-338c-4a4c-bb6f-41ba72d80861 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.664569] env[62600]: DEBUG oslo_vmware.api [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222298, 'name': CloneVM_Task} progress is 94%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.669091] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 787.669091] env[62600]: value = "task-1222302" [ 787.669091] env[62600]: _type = "Task" [ 787.669091] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.673259] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 787.673446] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 787.675167] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2fc83089-a5fa-4320-b17b-49df2515d6e3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.680401] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222302, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.683408] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Waiting for the task: (returnval){ [ 787.683408] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5296960d-1a93-2897-25b3-1678a61581d2" [ 787.683408] env[62600]: _type = "Task" [ 787.683408] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.691026] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5296960d-1a93-2897-25b3-1678a61581d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.780686] env[62600]: DEBUG oslo_concurrency.lockutils [None req-70b9cc24-35a8-43a7-8ee2-5156cd6522a5 tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Lock "e25c631d-3d1b-40d2-9fb1-a65431f991a9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.435s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.944880] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.512s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.945472] env[62600]: DEBUG nova.compute.manager [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 787.948524] env[62600]: DEBUG oslo_concurrency.lockutils [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 12.860s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.967321] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': task-1222300, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.039026] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222301, 'name': CreateVM_Task, 'duration_secs': 0.418159} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.039026] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 788.039026] env[62600]: DEBUG oslo_concurrency.lockutils [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.039026] env[62600]: DEBUG oslo_concurrency.lockutils [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.039026] env[62600]: DEBUG oslo_concurrency.lockutils [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 788.039026] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a87a76b-0a3d-4616-ae3c-5cb25f15888c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.048154] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 788.048154] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52036775-3217-6b12-8a95-5439ace15df9" [ 788.048154] env[62600]: _type = "Task" [ 788.048154] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.059543] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52036775-3217-6b12-8a95-5439ace15df9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.161097] env[62600]: DEBUG oslo_vmware.api [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222298, 'name': CloneVM_Task, 'duration_secs': 1.994796} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.161097] env[62600]: INFO nova.virt.vmwareapi.vmops [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Created linked-clone VM from snapshot [ 788.161097] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffd337a7-ec0e-4e35-8f16-d0b7929c9201 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.167266] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Uploading image 6f9a3616-6c62-4597-af46-3d4f504e1193 {{(pid=62600) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 788.178573] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222302, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.194942] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5296960d-1a93-2897-25b3-1678a61581d2, 'name': SearchDatastore_Task, 'duration_secs': 0.008796} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.197533] env[62600]: DEBUG oslo_vmware.rw_handles [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 788.197533] env[62600]: value = "vm-264237" [ 788.197533] env[62600]: _type = "VirtualMachine" [ 788.197533] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 788.198685] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-92c3be51-6d30-46af-abd3-557acaaf5593 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.200546] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-638aa90a-b2f3-408d-8f12-454779871614 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.206622] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Waiting for the task: (returnval){ [ 788.206622] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e4beaf-3df1-964a-ca14-da94633ba426" [ 788.206622] env[62600]: _type = "Task" [ 788.206622] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.208169] env[62600]: DEBUG oslo_vmware.rw_handles [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lease: (returnval){ [ 788.208169] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b3b140-dfe8-0358-a97b-2131b1ae2ae6" [ 788.208169] env[62600]: _type = "HttpNfcLease" [ 788.208169] env[62600]: } obtained for exporting VM: (result){ [ 788.208169] env[62600]: value = "vm-264237" [ 788.208169] env[62600]: _type = "VirtualMachine" [ 788.208169] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 788.208641] env[62600]: DEBUG oslo_vmware.api [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the lease: (returnval){ [ 788.208641] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b3b140-dfe8-0358-a97b-2131b1ae2ae6" [ 788.208641] env[62600]: _type = "HttpNfcLease" [ 788.208641] env[62600]: } to be ready. {{(pid=62600) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 788.219911] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 788.219911] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b3b140-dfe8-0358-a97b-2131b1ae2ae6" [ 788.219911] env[62600]: _type = "HttpNfcLease" [ 788.219911] env[62600]: } is initializing. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 788.223084] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e4beaf-3df1-964a-ca14-da94633ba426, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.283592] env[62600]: DEBUG nova.compute.manager [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 788.452572] env[62600]: DEBUG nova.compute.utils [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 788.457105] env[62600]: INFO nova.compute.claims [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 788.461321] env[62600]: DEBUG nova.compute.manager [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 788.461543] env[62600]: DEBUG nova.network.neutron [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 788.473569] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': task-1222300, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.559415] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52036775-3217-6b12-8a95-5439ace15df9, 'name': SearchDatastore_Task, 'duration_secs': 0.055015} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.560090] env[62600]: DEBUG oslo_concurrency.lockutils [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.560446] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 788.560565] env[62600]: DEBUG oslo_concurrency.lockutils [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.562854] env[62600]: DEBUG nova.policy [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4b2ae1429c03426a844cb2bd92d38256', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '583e67fe6a444511946ce9fac5c66420', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 788.679858] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222302, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.508925} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.682090] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc/b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 788.682090] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 788.682090] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b96489e5-ffd6-4f5a-a325-68ece9bf7759 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.687983] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 788.687983] env[62600]: value = "task-1222304" [ 788.687983] env[62600]: _type = "Task" [ 788.687983] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.700277] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222304, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.724562] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e4beaf-3df1-964a-ca14-da94633ba426, 'name': SearchDatastore_Task, 'duration_secs': 0.015141} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.727329] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.727628] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] dcbe8e2d-6ae9-465f-8394-3978ee61a15b/dcbe8e2d-6ae9-465f-8394-3978ee61a15b.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 788.728076] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 788.728076] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b3b140-dfe8-0358-a97b-2131b1ae2ae6" [ 788.728076] env[62600]: _type = "HttpNfcLease" [ 788.728076] env[62600]: } is ready. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 788.728430] env[62600]: DEBUG oslo_concurrency.lockutils [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.728673] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 788.728857] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1cb0fd03-450e-4c48-bce3-731f98f7c95e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.730842] env[62600]: DEBUG oslo_vmware.rw_handles [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 788.730842] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b3b140-dfe8-0358-a97b-2131b1ae2ae6" [ 788.730842] env[62600]: _type = "HttpNfcLease" [ 788.730842] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 788.731074] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bb7918b4-1abc-4306-ac6a-4bd045b4995f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.733486] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ebe054b-1216-4de9-898e-7d4942139fe1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.742550] env[62600]: DEBUG oslo_vmware.rw_handles [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5216eced-7788-1fb4-a39e-094004af51a1/disk-0.vmdk from lease info. {{(pid=62600) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 788.742861] env[62600]: DEBUG oslo_vmware.rw_handles [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5216eced-7788-1fb4-a39e-094004af51a1/disk-0.vmdk for reading. {{(pid=62600) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 788.746128] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 788.746316] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 788.747277] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Waiting for the task: (returnval){ [ 788.747277] env[62600]: value = "task-1222305" [ 788.747277] env[62600]: _type = "Task" [ 788.747277] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.748351] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e730918-9bc1-4972-908e-a3122833082b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.817106] env[62600]: DEBUG nova.compute.manager [req-ef53808c-beab-4f7b-aea9-79d514d28d07 req-5921df51-b0d6-4329-b6df-e8ea40669f10 service nova] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Received event network-changed-8ca7491d-af98-4a32-b8d9-63e6b57356ef {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 788.817343] env[62600]: DEBUG nova.compute.manager [req-ef53808c-beab-4f7b-aea9-79d514d28d07 req-5921df51-b0d6-4329-b6df-e8ea40669f10 service nova] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Refreshing instance network info cache due to event network-changed-8ca7491d-af98-4a32-b8d9-63e6b57356ef. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 788.817520] env[62600]: DEBUG oslo_concurrency.lockutils [req-ef53808c-beab-4f7b-aea9-79d514d28d07 req-5921df51-b0d6-4329-b6df-e8ea40669f10 service nova] Acquiring lock "refresh_cache-e25c631d-3d1b-40d2-9fb1-a65431f991a9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.817662] env[62600]: DEBUG oslo_concurrency.lockutils [req-ef53808c-beab-4f7b-aea9-79d514d28d07 req-5921df51-b0d6-4329-b6df-e8ea40669f10 service nova] Acquired lock "refresh_cache-e25c631d-3d1b-40d2-9fb1-a65431f991a9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.817820] env[62600]: DEBUG nova.network.neutron [req-ef53808c-beab-4f7b-aea9-79d514d28d07 req-5921df51-b0d6-4329-b6df-e8ea40669f10 service nova] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Refreshing network info cache for port 8ca7491d-af98-4a32-b8d9-63e6b57356ef {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 788.823576] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 788.823576] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c6b2a2-0b3a-8b56-9bd9-fd0233511bec" [ 788.823576] env[62600]: _type = "Task" [ 788.823576] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.830523] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': task-1222305, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.836809] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c6b2a2-0b3a-8b56-9bd9-fd0233511bec, 'name': SearchDatastore_Task, 'duration_secs': 0.016095} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.837611] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b237ae99-684d-42d3-b802-d1359b804847 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.840583] env[62600]: DEBUG oslo_concurrency.lockutils [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.843840] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 788.843840] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520ad9eb-6f51-4aa5-a90c-0e60b0206805" [ 788.843840] env[62600]: _type = "Task" [ 788.843840] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.849404] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9902b8bb-65a5-4796-9677-eee9d277d990 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.858832] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520ad9eb-6f51-4aa5-a90c-0e60b0206805, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.905458] env[62600]: DEBUG nova.network.neutron [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Successfully created port: cb1945e8-09df-4f1f-ba94-77e84b24f494 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 788.961861] env[62600]: DEBUG nova.compute.manager [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 788.966413] env[62600]: INFO nova.compute.resource_tracker [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Updating resource usage from migration a871b6f5-809b-4c75-97f7-3f72b2a3dc6a [ 788.978443] env[62600]: DEBUG oslo_vmware.api [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': task-1222300, 'name': PowerOnVM_Task, 'duration_secs': 1.041271} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.978766] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 788.978992] env[62600]: INFO nova.compute.manager [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Took 8.97 seconds to spawn the instance on the hypervisor. [ 788.979298] env[62600]: DEBUG nova.compute.manager [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 788.980107] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d40927d-5e39-4040-acf2-53df12ebadb6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.202356] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222304, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065825} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.202356] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 789.203114] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26fb943d-965b-4b31-b10d-3f757fb951a7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.226304] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc/b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 789.229954] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-069785d0-edb3-4b78-8427-5c171b3796de {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.259070] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 789.259070] env[62600]: value = "task-1222306" [ 789.259070] env[62600]: _type = "Task" [ 789.259070] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.276612] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': task-1222305, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.283602] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222306, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.361887] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520ad9eb-6f51-4aa5-a90c-0e60b0206805, 'name': SearchDatastore_Task, 'duration_secs': 0.013594} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.362876] env[62600]: DEBUG oslo_concurrency.lockutils [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.363321] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] d34c098d-a827-4dbd-96e4-e27d9d56b847/d34c098d-a827-4dbd-96e4-e27d9d56b847.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 789.366587] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-556eb237-1764-4e4f-80fb-e4b3d0c12a17 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.374680] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 789.374680] env[62600]: value = "task-1222307" [ 789.374680] env[62600]: _type = "Task" [ 789.374680] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.386733] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222307, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.501121] env[62600]: INFO nova.compute.manager [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Took 29.33 seconds to build instance. [ 789.549398] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b5dacb7-286f-41a7-9d20-2c23c00e2d9d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.561248] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abc541ac-3ad3-4395-a273-2b8ff23bba0d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.609798] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc0993d1-79fe-4f54-acbc-3d5464992f18 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.622850] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f4c6524-d242-4c19-a4cc-5cec7e4b993f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.646285] env[62600]: DEBUG nova.compute.provider_tree [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.723007] env[62600]: DEBUG nova.network.neutron [req-ef53808c-beab-4f7b-aea9-79d514d28d07 req-5921df51-b0d6-4329-b6df-e8ea40669f10 service nova] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Updated VIF entry in instance network info cache for port 8ca7491d-af98-4a32-b8d9-63e6b57356ef. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 789.723379] env[62600]: DEBUG nova.network.neutron [req-ef53808c-beab-4f7b-aea9-79d514d28d07 req-5921df51-b0d6-4329-b6df-e8ea40669f10 service nova] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Updating instance_info_cache with network_info: [{"id": "8ca7491d-af98-4a32-b8d9-63e6b57356ef", "address": "fa:16:3e:1a:e1:b5", "network": {"id": "f27f3669-f56f-4893-ae57-430b121f6309", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-417681146-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4e28048248a44a6fb0113b3f84a8e8be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "496faa4d-d874-449b-905e-328ddd60b31b", "external-id": "nsx-vlan-transportzone-391", "segmentation_id": 391, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ca7491d-af", "ovs_interfaceid": "8ca7491d-af98-4a32-b8d9-63e6b57356ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.762494] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': task-1222305, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.791777} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.762844] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] dcbe8e2d-6ae9-465f-8394-3978ee61a15b/dcbe8e2d-6ae9-465f-8394-3978ee61a15b.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 789.763164] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 789.763497] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f84f40ad-2c14-4772-8d6f-1b4b7462a269 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.774808] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222306, 'name': ReconfigVM_Task, 'duration_secs': 0.443024} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.776654] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Reconfigured VM instance instance-00000032 to attach disk [datastore2] b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc/b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 789.777801] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Waiting for the task: (returnval){ [ 789.777801] env[62600]: value = "task-1222308" [ 789.777801] env[62600]: _type = "Task" [ 789.777801] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.778246] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-78e26bb6-0541-45be-9c4d-0f13ad897c66 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.790432] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': task-1222308, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.792376] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 789.792376] env[62600]: value = "task-1222309" [ 789.792376] env[62600]: _type = "Task" [ 789.792376] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.801837] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222309, 'name': Rename_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.887442] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222307, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.979667] env[62600]: DEBUG nova.compute.manager [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 790.008221] env[62600]: DEBUG oslo_concurrency.lockutils [None req-dfbb7388-a8c2-4e03-971b-181ae06cdd57 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Lock "0a8840f1-7681-4aaa-9f3a-32b72c04c1c4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.026s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.008929] env[62600]: DEBUG nova.virt.hardware [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 790.009615] env[62600]: DEBUG nova.virt.hardware [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 790.009906] env[62600]: DEBUG nova.virt.hardware [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 790.010463] env[62600]: DEBUG nova.virt.hardware [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 790.010738] env[62600]: DEBUG nova.virt.hardware [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 790.011032] env[62600]: DEBUG nova.virt.hardware [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 790.011468] env[62600]: DEBUG nova.virt.hardware [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 790.011670] env[62600]: DEBUG nova.virt.hardware [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 790.011970] env[62600]: DEBUG nova.virt.hardware [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 790.012276] env[62600]: DEBUG nova.virt.hardware [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 790.012590] env[62600]: DEBUG nova.virt.hardware [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 790.017572] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c369a7d-7ed4-48b4-8871-c18d4b4c0ab6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.034449] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6429a975-8019-4dbb-a13d-a3e71468a99b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.150227] env[62600]: DEBUG nova.scheduler.client.report [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 790.228241] env[62600]: DEBUG oslo_concurrency.lockutils [req-ef53808c-beab-4f7b-aea9-79d514d28d07 req-5921df51-b0d6-4329-b6df-e8ea40669f10 service nova] Releasing lock "refresh_cache-e25c631d-3d1b-40d2-9fb1-a65431f991a9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.290904] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': task-1222308, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.111841} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.291347] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 790.292826] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fecd46ca-47cf-4c01-8bfc-27b3cb979200 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.323458] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] dcbe8e2d-6ae9-465f-8394-3978ee61a15b/dcbe8e2d-6ae9-465f-8394-3978ee61a15b.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 790.328180] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1dd907c-b36a-4a41-afdc-f65ebaa08b9e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.344536] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222309, 'name': Rename_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.350448] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Waiting for the task: (returnval){ [ 790.350448] env[62600]: value = "task-1222310" [ 790.350448] env[62600]: _type = "Task" [ 790.350448] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.358195] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': task-1222310, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.387239] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222307, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.673945} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.387537] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] d34c098d-a827-4dbd-96e4-e27d9d56b847/d34c098d-a827-4dbd-96e4-e27d9d56b847.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 790.387756] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 790.388017] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2a9222ee-b6b3-43af-82e6-1f577f0a9c5f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.394702] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 790.394702] env[62600]: value = "task-1222311" [ 790.394702] env[62600]: _type = "Task" [ 790.394702] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.403541] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222311, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.524029] env[62600]: DEBUG nova.compute.manager [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 790.573730] env[62600]: DEBUG nova.compute.manager [req-227a25b9-6616-4521-bc2d-e8f39d592da0 req-3f63dfda-955a-4d4f-ad9c-da5983a3ccc5 service nova] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Received event network-vif-plugged-cb1945e8-09df-4f1f-ba94-77e84b24f494 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 790.574025] env[62600]: DEBUG oslo_concurrency.lockutils [req-227a25b9-6616-4521-bc2d-e8f39d592da0 req-3f63dfda-955a-4d4f-ad9c-da5983a3ccc5 service nova] Acquiring lock "92bd8b10-2034-49d4-bf38-0edc7e1e6f7d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.574200] env[62600]: DEBUG oslo_concurrency.lockutils [req-227a25b9-6616-4521-bc2d-e8f39d592da0 req-3f63dfda-955a-4d4f-ad9c-da5983a3ccc5 service nova] Lock "92bd8b10-2034-49d4-bf38-0edc7e1e6f7d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.574378] env[62600]: DEBUG oslo_concurrency.lockutils [req-227a25b9-6616-4521-bc2d-e8f39d592da0 req-3f63dfda-955a-4d4f-ad9c-da5983a3ccc5 service nova] Lock "92bd8b10-2034-49d4-bf38-0edc7e1e6f7d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.574567] env[62600]: DEBUG nova.compute.manager [req-227a25b9-6616-4521-bc2d-e8f39d592da0 req-3f63dfda-955a-4d4f-ad9c-da5983a3ccc5 service nova] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] No waiting events found dispatching network-vif-plugged-cb1945e8-09df-4f1f-ba94-77e84b24f494 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 790.574732] env[62600]: WARNING nova.compute.manager [req-227a25b9-6616-4521-bc2d-e8f39d592da0 req-3f63dfda-955a-4d4f-ad9c-da5983a3ccc5 service nova] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Received unexpected event network-vif-plugged-cb1945e8-09df-4f1f-ba94-77e84b24f494 for instance with vm_state building and task_state spawning. [ 790.655795] env[62600]: DEBUG oslo_concurrency.lockutils [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.707s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.656043] env[62600]: INFO nova.compute.manager [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Migrating [ 790.656337] env[62600]: DEBUG oslo_concurrency.lockutils [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.656571] env[62600]: DEBUG oslo_concurrency.lockutils [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.658140] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.973s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.660275] env[62600]: INFO nova.compute.claims [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 790.729036] env[62600]: DEBUG nova.network.neutron [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Successfully updated port: cb1945e8-09df-4f1f-ba94-77e84b24f494 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 790.807834] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222309, 'name': Rename_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.816984] env[62600]: DEBUG nova.compute.manager [req-cfe16455-2e66-42fa-90fb-326ad70606ee req-c87d7594-137a-454e-b156-bc1c7eaf467e service nova] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Received event network-changed-194093a1-33d0-42f8-8255-db5802300692 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 790.817226] env[62600]: DEBUG nova.compute.manager [req-cfe16455-2e66-42fa-90fb-326ad70606ee req-c87d7594-137a-454e-b156-bc1c7eaf467e service nova] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Refreshing instance network info cache due to event network-changed-194093a1-33d0-42f8-8255-db5802300692. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 790.817456] env[62600]: DEBUG oslo_concurrency.lockutils [req-cfe16455-2e66-42fa-90fb-326ad70606ee req-c87d7594-137a-454e-b156-bc1c7eaf467e service nova] Acquiring lock "refresh_cache-0a8840f1-7681-4aaa-9f3a-32b72c04c1c4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.817596] env[62600]: DEBUG oslo_concurrency.lockutils [req-cfe16455-2e66-42fa-90fb-326ad70606ee req-c87d7594-137a-454e-b156-bc1c7eaf467e service nova] Acquired lock "refresh_cache-0a8840f1-7681-4aaa-9f3a-32b72c04c1c4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.817758] env[62600]: DEBUG nova.network.neutron [req-cfe16455-2e66-42fa-90fb-326ad70606ee req-c87d7594-137a-454e-b156-bc1c7eaf467e service nova] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Refreshing network info cache for port 194093a1-33d0-42f8-8255-db5802300692 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 790.864956] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': task-1222310, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.905752] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222311, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070519} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.908255] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 790.908255] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e06ce262-7ae5-435a-a379-90f7f37be1f7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.931020] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] d34c098d-a827-4dbd-96e4-e27d9d56b847/d34c098d-a827-4dbd-96e4-e27d9d56b847.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 790.931020] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3fb7cba0-9272-4604-8bf6-fe621792e1d6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.950890] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 790.950890] env[62600]: value = "task-1222312" [ 790.950890] env[62600]: _type = "Task" [ 790.950890] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.960407] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222312, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.049762] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.170023] env[62600]: INFO nova.compute.rpcapi [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 791.170023] env[62600]: DEBUG oslo_concurrency.lockutils [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.234228] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquiring lock "refresh_cache-92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.234404] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquired lock "refresh_cache-92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.234653] env[62600]: DEBUG nova.network.neutron [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 791.307992] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222309, 'name': Rename_Task, 'duration_secs': 1.205744} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.308639] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 791.309427] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aae81ee3-747b-4660-a74d-41275da1615a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.318234] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 791.318234] env[62600]: value = "task-1222313" [ 791.318234] env[62600]: _type = "Task" [ 791.318234] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.328835] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222313, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.361526] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': task-1222310, 'name': ReconfigVM_Task, 'duration_secs': 0.603363} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.361915] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Reconfigured VM instance instance-00000031 to attach disk [datastore2] dcbe8e2d-6ae9-465f-8394-3978ee61a15b/dcbe8e2d-6ae9-465f-8394-3978ee61a15b.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 791.362587] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93db0cfa-ddfe-4e78-841f-46dbca14ef21 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.369161] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Waiting for the task: (returnval){ [ 791.369161] env[62600]: value = "task-1222314" [ 791.369161] env[62600]: _type = "Task" [ 791.369161] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.377866] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': task-1222314, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.464699] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222312, 'name': ReconfigVM_Task, 'duration_secs': 0.388274} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.465127] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Reconfigured VM instance instance-00000033 to attach disk [datastore2] d34c098d-a827-4dbd-96e4-e27d9d56b847/d34c098d-a827-4dbd-96e4-e27d9d56b847.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 791.465848] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e640e72f-b92f-4485-8920-36719c564d09 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.473026] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 791.473026] env[62600]: value = "task-1222315" [ 791.473026] env[62600]: _type = "Task" [ 791.473026] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.481734] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222315, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.690812] env[62600]: DEBUG oslo_concurrency.lockutils [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "refresh_cache-de4f9304-3357-4eaa-9c94-fe28bc554086" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.691013] env[62600]: DEBUG oslo_concurrency.lockutils [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquired lock "refresh_cache-de4f9304-3357-4eaa-9c94-fe28bc554086" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.691203] env[62600]: DEBUG nova.network.neutron [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 791.744142] env[62600]: DEBUG nova.network.neutron [req-cfe16455-2e66-42fa-90fb-326ad70606ee req-c87d7594-137a-454e-b156-bc1c7eaf467e service nova] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Updated VIF entry in instance network info cache for port 194093a1-33d0-42f8-8255-db5802300692. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 791.744522] env[62600]: DEBUG nova.network.neutron [req-cfe16455-2e66-42fa-90fb-326ad70606ee req-c87d7594-137a-454e-b156-bc1c7eaf467e service nova] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Updating instance_info_cache with network_info: [{"id": "194093a1-33d0-42f8-8255-db5802300692", "address": "fa:16:3e:4e:df:59", "network": {"id": "884e4182-92a8-4da3-80bc-b98a3fb9b0ae", "bridge": "br-int", "label": "tempest-ServersTestJSON-883977721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.137", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "880c2a040cca4ee685451a87f8dcab28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap194093a1-33", "ovs_interfaceid": "194093a1-33d0-42f8-8255-db5802300692", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.788188] env[62600]: DEBUG nova.network.neutron [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 791.828760] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222313, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.880212] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': task-1222314, 'name': Rename_Task, 'duration_secs': 0.181454} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.880212] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 791.882517] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-76fe1f8a-508b-490c-8b03-70790264690a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.888717] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Waiting for the task: (returnval){ [ 791.888717] env[62600]: value = "task-1222316" [ 791.888717] env[62600]: _type = "Task" [ 791.888717] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.895831] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': task-1222316, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.950727] env[62600]: DEBUG nova.network.neutron [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Updating instance_info_cache with network_info: [{"id": "cb1945e8-09df-4f1f-ba94-77e84b24f494", "address": "fa:16:3e:63:98:19", "network": {"id": "f9decfe0-071c-4ff4-b65a-ea34be9fc487", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-439991991-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "583e67fe6a444511946ce9fac5c66420", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1ce8361b-fd8e-4971-a37f-b84a4f77db19", "external-id": "nsx-vlan-transportzone-255", "segmentation_id": 255, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb1945e8-09", "ovs_interfaceid": "cb1945e8-09df-4f1f-ba94-77e84b24f494", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.983190] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222315, 'name': Rename_Task, 'duration_secs': 0.19699} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.985729] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 791.986400] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6fab9643-3eab-48a2-9702-185063232f23 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.992167] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 791.992167] env[62600]: value = "task-1222317" [ 791.992167] env[62600]: _type = "Task" [ 791.992167] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.002272] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222317, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.047842] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64d312c7-4a29-4678-9b6f-87145a898c7d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.055303] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a127c77-f371-4509-82f1-aea9a2b24694 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.088806] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130370fb-7426-46ea-adaf-b95c6fa6a835 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.096308] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1427a04-df3c-4b4a-8dcd-7e552b241af0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.110260] env[62600]: DEBUG nova.compute.provider_tree [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.248691] env[62600]: DEBUG oslo_concurrency.lockutils [req-cfe16455-2e66-42fa-90fb-326ad70606ee req-c87d7594-137a-454e-b156-bc1c7eaf467e service nova] Releasing lock "refresh_cache-0a8840f1-7681-4aaa-9f3a-32b72c04c1c4" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.328931] env[62600]: DEBUG oslo_vmware.api [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222313, 'name': PowerOnVM_Task, 'duration_secs': 0.54455} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.329219] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 792.329424] env[62600]: INFO nova.compute.manager [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Took 7.38 seconds to spawn the instance on the hypervisor. [ 792.329718] env[62600]: DEBUG nova.compute.manager [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 792.330472] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86c203c4-38b6-4baf-a7db-f2ca546bb50b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.397928] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': task-1222316, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.412074] env[62600]: DEBUG nova.network.neutron [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Updating instance_info_cache with network_info: [{"id": "fe02d314-372f-43fb-8d7c-cfc33c9e74bd", "address": "fa:16:3e:86:22:ec", "network": {"id": "d946b94e-9827-4c22-af05-072fcaa3d0e6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.235", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ec39f2547f624ed0b240d084b949694a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe02d314-37", "ovs_interfaceid": "fe02d314-372f-43fb-8d7c-cfc33c9e74bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.452513] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Releasing lock "refresh_cache-92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.452878] env[62600]: DEBUG nova.compute.manager [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Instance network_info: |[{"id": "cb1945e8-09df-4f1f-ba94-77e84b24f494", "address": "fa:16:3e:63:98:19", "network": {"id": "f9decfe0-071c-4ff4-b65a-ea34be9fc487", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-439991991-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "583e67fe6a444511946ce9fac5c66420", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1ce8361b-fd8e-4971-a37f-b84a4f77db19", "external-id": "nsx-vlan-transportzone-255", "segmentation_id": 255, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb1945e8-09", "ovs_interfaceid": "cb1945e8-09df-4f1f-ba94-77e84b24f494", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 792.453410] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:98:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1ce8361b-fd8e-4971-a37f-b84a4f77db19', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cb1945e8-09df-4f1f-ba94-77e84b24f494', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 792.461216] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Creating folder: Project (583e67fe6a444511946ce9fac5c66420). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 792.462324] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bcdb067c-bb14-4381-a626-6ee2c59b4335 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.472155] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Created folder: Project (583e67fe6a444511946ce9fac5c66420) in parent group-v264198. [ 792.472369] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Creating folder: Instances. Parent ref: group-v264239. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 792.472713] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ec330637-ae17-4ec5-b150-bb1d108d3319 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.481800] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Created folder: Instances in parent group-v264239. [ 792.481800] env[62600]: DEBUG oslo.service.loopingcall [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 792.481953] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 792.482615] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d6b94d72-4891-4368-8729-40e8ec311eba {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.505447] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222317, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.506611] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 792.506611] env[62600]: value = "task-1222320" [ 792.506611] env[62600]: _type = "Task" [ 792.506611] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.513685] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222320, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.603089] env[62600]: DEBUG nova.compute.manager [req-bd70bdab-5b71-4675-9076-7de1d5b76358 req-32515d72-afc8-451e-9194-4deb8d6fb63c service nova] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Received event network-changed-cb1945e8-09df-4f1f-ba94-77e84b24f494 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 792.603330] env[62600]: DEBUG nova.compute.manager [req-bd70bdab-5b71-4675-9076-7de1d5b76358 req-32515d72-afc8-451e-9194-4deb8d6fb63c service nova] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Refreshing instance network info cache due to event network-changed-cb1945e8-09df-4f1f-ba94-77e84b24f494. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 792.603559] env[62600]: DEBUG oslo_concurrency.lockutils [req-bd70bdab-5b71-4675-9076-7de1d5b76358 req-32515d72-afc8-451e-9194-4deb8d6fb63c service nova] Acquiring lock "refresh_cache-92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.603708] env[62600]: DEBUG oslo_concurrency.lockutils [req-bd70bdab-5b71-4675-9076-7de1d5b76358 req-32515d72-afc8-451e-9194-4deb8d6fb63c service nova] Acquired lock "refresh_cache-92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.603939] env[62600]: DEBUG nova.network.neutron [req-bd70bdab-5b71-4675-9076-7de1d5b76358 req-32515d72-afc8-451e-9194-4deb8d6fb63c service nova] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Refreshing network info cache for port cb1945e8-09df-4f1f-ba94-77e84b24f494 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 792.612934] env[62600]: DEBUG nova.scheduler.client.report [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 792.854163] env[62600]: INFO nova.compute.manager [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Took 28.88 seconds to build instance. [ 792.901823] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': task-1222316, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.915722] env[62600]: DEBUG oslo_concurrency.lockutils [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Releasing lock "refresh_cache-de4f9304-3357-4eaa-9c94-fe28bc554086" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.010055] env[62600]: DEBUG oslo_vmware.api [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222317, 'name': PowerOnVM_Task, 'duration_secs': 0.996992} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.013102] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 793.013513] env[62600]: INFO nova.compute.manager [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Took 5.56 seconds to spawn the instance on the hypervisor. [ 793.013734] env[62600]: DEBUG nova.compute.manager [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 793.014863] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1017304-fce8-4ed3-947a-837ac747b6b7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.022222] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222320, 'name': CreateVM_Task, 'duration_secs': 0.412302} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.024208] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 793.027429] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.027543] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.028209] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 793.028638] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87cf2b97-fa58-4dbc-9a9d-8020ef02238a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.033445] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Waiting for the task: (returnval){ [ 793.033445] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52698102-ef80-3aaf-7735-c5f329312aeb" [ 793.033445] env[62600]: _type = "Task" [ 793.033445] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.041637] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52698102-ef80-3aaf-7735-c5f329312aeb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.118365] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.119136] env[62600]: DEBUG nova.compute.manager [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 793.121967] env[62600]: DEBUG oslo_concurrency.lockutils [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.678s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.124042] env[62600]: INFO nova.compute.claims [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 793.330120] env[62600]: DEBUG nova.network.neutron [req-bd70bdab-5b71-4675-9076-7de1d5b76358 req-32515d72-afc8-451e-9194-4deb8d6fb63c service nova] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Updated VIF entry in instance network info cache for port cb1945e8-09df-4f1f-ba94-77e84b24f494. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 793.330523] env[62600]: DEBUG nova.network.neutron [req-bd70bdab-5b71-4675-9076-7de1d5b76358 req-32515d72-afc8-451e-9194-4deb8d6fb63c service nova] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Updating instance_info_cache with network_info: [{"id": "cb1945e8-09df-4f1f-ba94-77e84b24f494", "address": "fa:16:3e:63:98:19", "network": {"id": "f9decfe0-071c-4ff4-b65a-ea34be9fc487", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-439991991-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "583e67fe6a444511946ce9fac5c66420", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1ce8361b-fd8e-4971-a37f-b84a4f77db19", "external-id": "nsx-vlan-transportzone-255", "segmentation_id": 255, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb1945e8-09", "ovs_interfaceid": "cb1945e8-09df-4f1f-ba94-77e84b24f494", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.356410] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b3680f40-d3ce-49cb-a533-5190dc97cace tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.103s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.399015] env[62600]: DEBUG oslo_vmware.api [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': task-1222316, 'name': PowerOnVM_Task, 'duration_secs': 1.022362} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.399509] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 793.399803] env[62600]: INFO nova.compute.manager [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Took 10.92 seconds to spawn the instance on the hypervisor. [ 793.399961] env[62600]: DEBUG nova.compute.manager [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 793.400764] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-437b9f14-bd41-4d4d-99b9-46caf9b475ac {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.540329] env[62600]: INFO nova.compute.manager [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Took 22.73 seconds to build instance. [ 793.547821] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52698102-ef80-3aaf-7735-c5f329312aeb, 'name': SearchDatastore_Task, 'duration_secs': 0.017236} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.548219] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.549477] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 793.549477] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.549477] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.549477] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 793.549884] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-316215e5-7b50-4b06-b650-f8ee7f4a93d6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.559339] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 793.559550] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 793.560312] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a513f89-94f2-403c-b199-4e1c349ce742 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.566382] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Waiting for the task: (returnval){ [ 793.566382] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52fff9c7-2b9d-1338-41be-2d3698b8920a" [ 793.566382] env[62600]: _type = "Task" [ 793.566382] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.574399] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52fff9c7-2b9d-1338-41be-2d3698b8920a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.629505] env[62600]: DEBUG nova.compute.utils [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 793.633106] env[62600]: DEBUG nova.compute.manager [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 793.633349] env[62600]: DEBUG nova.network.neutron [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 793.672165] env[62600]: DEBUG nova.policy [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06fe44327feb4673a1489d4d040be8ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '03b37f47579d415384c4cd92bef5ca00', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 793.833818] env[62600]: DEBUG oslo_concurrency.lockutils [req-bd70bdab-5b71-4675-9076-7de1d5b76358 req-32515d72-afc8-451e-9194-4deb8d6fb63c service nova] Releasing lock "refresh_cache-92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.858857] env[62600]: DEBUG nova.compute.manager [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 793.920719] env[62600]: INFO nova.compute.manager [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Took 31.87 seconds to build instance. [ 793.961385] env[62600]: DEBUG nova.network.neutron [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Successfully created port: ab8eb6fb-6c95-4452-9f98-963a04b0d64d {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 794.042958] env[62600]: DEBUG oslo_concurrency.lockutils [None req-49551440-9247-4e94-ab59-c67bf9d36e0e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "d34c098d-a827-4dbd-96e4-e27d9d56b847" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.260s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.083258] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52fff9c7-2b9d-1338-41be-2d3698b8920a, 'name': SearchDatastore_Task, 'duration_secs': 0.013159} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.084111] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-adc4605d-5437-4854-b9e5-eff43cdc48f0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.089296] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Waiting for the task: (returnval){ [ 794.089296] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52839185-1e99-b70b-c18f-ac3dbc42016c" [ 794.089296] env[62600]: _type = "Task" [ 794.089296] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.097725] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52839185-1e99-b70b-c18f-ac3dbc42016c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.134704] env[62600]: DEBUG nova.compute.manager [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 794.292172] env[62600]: INFO nova.compute.manager [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Rebuilding instance [ 794.333221] env[62600]: DEBUG nova.compute.manager [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 794.333705] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b0e611-c86b-4cd0-9d46-1457b9a5abb9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.381818] env[62600]: DEBUG oslo_concurrency.lockutils [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.422247] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c35844df-eeae-4dcd-83c8-05bfe5f8e681 tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Lock "dcbe8e2d-6ae9-465f-8394-3978ee61a15b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.327s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.433974] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6c943af-c3d0-49c1-8426-394680083f8c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.454632] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Updating instance 'de4f9304-3357-4eaa-9c94-fe28bc554086' progress to 0 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 794.538327] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e00058f-e431-4842-a80a-b1d6b12c518a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.545782] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af51ac57-520b-4886-9698-c6970f5e9d9b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.549333] env[62600]: DEBUG nova.compute.manager [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 794.580203] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcedd9ae-a367-408f-bddf-139ba24d9f77 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.590409] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1776187b-a8f5-4e2a-8d1f-26054d1eea8c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.609267] env[62600]: DEBUG nova.compute.provider_tree [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.614424] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52839185-1e99-b70b-c18f-ac3dbc42016c, 'name': SearchDatastore_Task, 'duration_secs': 0.015072} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.615123] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.615448] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d/92bd8b10-2034-49d4-bf38-0edc7e1e6f7d.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 794.615740] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-be4290de-12a4-4299-a329-5d4cf6652068 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.623705] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Waiting for the task: (returnval){ [ 794.623705] env[62600]: value = "task-1222321" [ 794.623705] env[62600]: _type = "Task" [ 794.623705] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.633623] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': task-1222321, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.717057] env[62600]: DEBUG oslo_concurrency.lockutils [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Acquiring lock "dcbe8e2d-6ae9-465f-8394-3978ee61a15b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.719064] env[62600]: DEBUG oslo_concurrency.lockutils [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Lock "dcbe8e2d-6ae9-465f-8394-3978ee61a15b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.719064] env[62600]: DEBUG oslo_concurrency.lockutils [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Acquiring lock "dcbe8e2d-6ae9-465f-8394-3978ee61a15b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.719064] env[62600]: DEBUG oslo_concurrency.lockutils [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Lock "dcbe8e2d-6ae9-465f-8394-3978ee61a15b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.719064] env[62600]: DEBUG oslo_concurrency.lockutils [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Lock "dcbe8e2d-6ae9-465f-8394-3978ee61a15b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.721439] env[62600]: INFO nova.compute.manager [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Terminating instance [ 794.723827] env[62600]: DEBUG nova.compute.manager [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 794.724178] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 794.725185] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6285d43-a062-47c5-80a2-f8cd630e8f3c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.732728] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 794.733036] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5be31270-e2ed-4901-8ceb-9102a99e7904 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.739557] env[62600]: DEBUG oslo_vmware.api [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Waiting for the task: (returnval){ [ 794.739557] env[62600]: value = "task-1222322" [ 794.739557] env[62600]: _type = "Task" [ 794.739557] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.748512] env[62600]: DEBUG oslo_vmware.api [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': task-1222322, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.848337] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 794.848800] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-23efc1e2-8ee5-4876-8339-da7b281a5bca {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.857010] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 794.857010] env[62600]: value = "task-1222323" [ 794.857010] env[62600]: _type = "Task" [ 794.857010] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.871015] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222323, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.925947] env[62600]: DEBUG nova.compute.manager [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 794.963602] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 794.964576] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7183cd8b-a2c9-42c1-832c-8c5ccc2cc2f6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.974449] env[62600]: DEBUG oslo_vmware.api [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 794.974449] env[62600]: value = "task-1222324" [ 794.974449] env[62600]: _type = "Task" [ 794.974449] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.987333] env[62600]: DEBUG oslo_vmware.api [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222324, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.077852] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.117122] env[62600]: DEBUG nova.scheduler.client.report [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.134692] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': task-1222321, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.147690] env[62600]: DEBUG nova.compute.manager [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 795.175927] env[62600]: DEBUG nova.virt.hardware [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 795.176352] env[62600]: DEBUG nova.virt.hardware [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 795.176752] env[62600]: DEBUG nova.virt.hardware [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 795.177111] env[62600]: DEBUG nova.virt.hardware [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 795.177642] env[62600]: DEBUG nova.virt.hardware [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 795.177764] env[62600]: DEBUG nova.virt.hardware [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 795.177931] env[62600]: DEBUG nova.virt.hardware [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 795.178121] env[62600]: DEBUG nova.virt.hardware [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 795.178300] env[62600]: DEBUG nova.virt.hardware [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 795.178472] env[62600]: DEBUG nova.virt.hardware [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 795.178734] env[62600]: DEBUG nova.virt.hardware [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 795.179971] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9566ee1f-b0c6-4ed4-9401-4e74db93bd24 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.190193] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ef5b247-4fe4-44fd-9e06-2b921bf9884f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.249915] env[62600]: DEBUG oslo_vmware.api [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': task-1222322, 'name': PowerOffVM_Task, 'duration_secs': 0.225749} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.250241] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 795.250453] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 795.250723] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-786e7542-0784-4c51-b99e-a9b94f766c0b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.367555] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222323, 'name': PowerOffVM_Task, 'duration_secs': 0.154434} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.367555] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 795.367555] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 795.368386] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b69f67e-da78-4e10-9b33-032a9d7e5e4c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.375059] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 795.375315] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e9b6ecb-4e12-4ff2-a040-ba6e59692581 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.402958] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 795.403232] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 795.403419] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Deleting the datastore file [datastore2] d34c098d-a827-4dbd-96e4-e27d9d56b847 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 795.403689] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ee5eb87c-dd73-410b-9c7a-bd9d0631152e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.410171] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 795.410171] env[62600]: value = "task-1222327" [ 795.410171] env[62600]: _type = "Task" [ 795.410171] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.419228] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222327, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.446303] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.459172] env[62600]: DEBUG nova.compute.manager [req-798392b3-4d62-470b-af1a-32cecacf067f req-7864a3ea-94e5-481d-97f0-5643f93a3e76 service nova] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Received event network-vif-plugged-ab8eb6fb-6c95-4452-9f98-963a04b0d64d {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 795.459468] env[62600]: DEBUG oslo_concurrency.lockutils [req-798392b3-4d62-470b-af1a-32cecacf067f req-7864a3ea-94e5-481d-97f0-5643f93a3e76 service nova] Acquiring lock "734e8a2b-faaa-4165-b0e9-69eeb3699b2d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.459694] env[62600]: DEBUG oslo_concurrency.lockutils [req-798392b3-4d62-470b-af1a-32cecacf067f req-7864a3ea-94e5-481d-97f0-5643f93a3e76 service nova] Lock "734e8a2b-faaa-4165-b0e9-69eeb3699b2d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.459890] env[62600]: DEBUG oslo_concurrency.lockutils [req-798392b3-4d62-470b-af1a-32cecacf067f req-7864a3ea-94e5-481d-97f0-5643f93a3e76 service nova] Lock "734e8a2b-faaa-4165-b0e9-69eeb3699b2d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.460221] env[62600]: DEBUG nova.compute.manager [req-798392b3-4d62-470b-af1a-32cecacf067f req-7864a3ea-94e5-481d-97f0-5643f93a3e76 service nova] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] No waiting events found dispatching network-vif-plugged-ab8eb6fb-6c95-4452-9f98-963a04b0d64d {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 795.460318] env[62600]: WARNING nova.compute.manager [req-798392b3-4d62-470b-af1a-32cecacf067f req-7864a3ea-94e5-481d-97f0-5643f93a3e76 service nova] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Received unexpected event network-vif-plugged-ab8eb6fb-6c95-4452-9f98-963a04b0d64d for instance with vm_state building and task_state spawning. [ 795.488496] env[62600]: DEBUG oslo_vmware.api [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222324, 'name': PowerOffVM_Task, 'duration_secs': 0.222697} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.488496] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 795.488496] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Updating instance 'de4f9304-3357-4eaa-9c94-fe28bc554086' progress to 17 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 795.573529] env[62600]: DEBUG nova.network.neutron [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Successfully updated port: ab8eb6fb-6c95-4452-9f98-963a04b0d64d {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 795.622895] env[62600]: DEBUG oslo_concurrency.lockutils [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.501s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.623486] env[62600]: DEBUG nova.compute.manager [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 795.628106] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.294s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.628106] env[62600]: INFO nova.compute.claims [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 795.641524] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': task-1222321, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.551109} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.642744] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d/92bd8b10-2034-49d4-bf38-0edc7e1e6f7d.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 795.642744] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 795.642744] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9819d1ee-6611-461f-be79-54f85c38c5ff {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.652650] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Waiting for the task: (returnval){ [ 795.652650] env[62600]: value = "task-1222328" [ 795.652650] env[62600]: _type = "Task" [ 795.652650] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.662103] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': task-1222328, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.919848] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222327, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.317089} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.920174] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 795.920317] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 795.920502] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 795.994044] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 795.994323] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 795.994484] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 795.994672] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 795.994825] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 795.994977] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 795.995206] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 795.995370] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 795.995540] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 795.995712] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 795.995891] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 796.000986] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2e28720-a1d0-41be-bbc1-35b6fd6d5efc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.023210] env[62600]: DEBUG oslo_vmware.api [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 796.023210] env[62600]: value = "task-1222329" [ 796.023210] env[62600]: _type = "Task" [ 796.023210] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.031462] env[62600]: DEBUG oslo_vmware.api [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222329, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.078387] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Acquiring lock "refresh_cache-734e8a2b-faaa-4165-b0e9-69eeb3699b2d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.078609] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Acquired lock "refresh_cache-734e8a2b-faaa-4165-b0e9-69eeb3699b2d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.078775] env[62600]: DEBUG nova.network.neutron [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 796.132948] env[62600]: DEBUG nova.compute.utils [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 796.134553] env[62600]: DEBUG nova.compute.manager [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 796.134772] env[62600]: DEBUG nova.network.neutron [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 796.163070] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': task-1222328, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076544} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.163389] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 796.164342] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-956e3cf9-acf0-4caa-8851-f0ef59487665 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.188816] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d/92bd8b10-2034-49d4-bf38-0edc7e1e6f7d.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 796.191128] env[62600]: DEBUG nova.policy [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cecd66b578294ec9b01b662771297650', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cdf4ad7a2c214abd99b6e24550ac157e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 796.194029] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b8904ce-56de-4697-afce-583928895564 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.216067] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Waiting for the task: (returnval){ [ 796.216067] env[62600]: value = "task-1222330" [ 796.216067] env[62600]: _type = "Task" [ 796.216067] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.224874] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': task-1222330, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.503706] env[62600]: DEBUG nova.network.neutron [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Successfully created port: 73800cd5-a9c3-41ae-b4fe-fbb4d32081b7 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 796.534279] env[62600]: DEBUG oslo_vmware.api [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222329, 'name': ReconfigVM_Task, 'duration_secs': 0.244551} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.534597] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Updating instance 'de4f9304-3357-4eaa-9c94-fe28bc554086' progress to 33 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 796.629136] env[62600]: DEBUG nova.network.neutron [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 796.637782] env[62600]: DEBUG nova.compute.manager [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 796.696192] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 796.697465] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 796.697465] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Deleting the datastore file [datastore2] dcbe8e2d-6ae9-465f-8394-3978ee61a15b {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 796.697465] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eb9b8b77-2775-4d04-b801-a7492d7f9f07 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.709441] env[62600]: DEBUG oslo_vmware.api [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Waiting for the task: (returnval){ [ 796.709441] env[62600]: value = "task-1222331" [ 796.709441] env[62600]: _type = "Task" [ 796.709441] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.718238] env[62600]: DEBUG oslo_vmware.api [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': task-1222331, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.733552] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': task-1222330, 'name': ReconfigVM_Task, 'duration_secs': 0.395211} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.733954] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Reconfigured VM instance instance-00000034 to attach disk [datastore1] 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d/92bd8b10-2034-49d4-bf38-0edc7e1e6f7d.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 796.734648] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4b521dd8-4d6f-409d-8364-e6330bf23d1e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.745078] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Waiting for the task: (returnval){ [ 796.745078] env[62600]: value = "task-1222332" [ 796.745078] env[62600]: _type = "Task" [ 796.745078] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.761400] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': task-1222332, 'name': Rename_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.960538] env[62600]: DEBUG nova.network.neutron [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Updating instance_info_cache with network_info: [{"id": "ab8eb6fb-6c95-4452-9f98-963a04b0d64d", "address": "fa:16:3e:81:b6:6b", "network": {"id": "2edc47d5-8cf8-4d69-8630-3fe0e39e1c06", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-386075803-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03b37f47579d415384c4cd92bef5ca00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab8eb6fb-6c", "ovs_interfaceid": "ab8eb6fb-6c95-4452-9f98-963a04b0d64d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.980073] env[62600]: DEBUG nova.virt.hardware [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 796.980340] env[62600]: DEBUG nova.virt.hardware [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 796.981908] env[62600]: DEBUG nova.virt.hardware [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 796.981908] env[62600]: DEBUG nova.virt.hardware [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 796.981908] env[62600]: DEBUG nova.virt.hardware [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 796.981908] env[62600]: DEBUG nova.virt.hardware [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 796.981908] env[62600]: DEBUG nova.virt.hardware [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 796.982153] env[62600]: DEBUG nova.virt.hardware [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 796.982153] env[62600]: DEBUG nova.virt.hardware [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 796.982153] env[62600]: DEBUG nova.virt.hardware [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 796.982153] env[62600]: DEBUG nova.virt.hardware [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 796.983073] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30265343-162f-466c-879e-7827a1de1bf7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.995979] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73ebb768-c352-463f-a64d-666d1679069a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.013074] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Instance VIF info [] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 797.019195] env[62600]: DEBUG oslo.service.loopingcall [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.021948] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 797.022459] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-800f8504-58e1-42a5-b36e-663ccbce51d3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.042117] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T12:01:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='99810421-2db4-4272-82bd-17ba5175b496',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-2077855957',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 797.042844] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 797.042844] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 797.042844] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 797.043054] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 797.043145] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 797.043336] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 797.043660] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 797.043716] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 797.044120] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 797.044120] env[62600]: DEBUG nova.virt.hardware [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 797.049722] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Reconfiguring VM instance instance-0000002b to detach disk 2000 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 797.054365] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e82465d-7ecb-4acd-b6ef-932d35f5d8f0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.070764] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 797.070764] env[62600]: value = "task-1222333" [ 797.070764] env[62600]: _type = "Task" [ 797.070764] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.078228] env[62600]: DEBUG oslo_vmware.api [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 797.078228] env[62600]: value = "task-1222334" [ 797.078228] env[62600]: _type = "Task" [ 797.078228] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.078228] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222333, 'name': CreateVM_Task} progress is 15%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.091678] env[62600]: DEBUG oslo_vmware.api [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222334, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.149078] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbaa86be-d822-4ce2-8f6e-b6c9ab6cc9fa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.159218] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dde0601-9d41-4a1d-a7e9-15fb52818bfe {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.955308] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Releasing lock "refresh_cache-734e8a2b-faaa-4165-b0e9-69eeb3699b2d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.955742] env[62600]: DEBUG nova.compute.manager [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Instance network_info: |[{"id": "ab8eb6fb-6c95-4452-9f98-963a04b0d64d", "address": "fa:16:3e:81:b6:6b", "network": {"id": "2edc47d5-8cf8-4d69-8630-3fe0e39e1c06", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-386075803-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03b37f47579d415384c4cd92bef5ca00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab8eb6fb-6c", "ovs_interfaceid": "ab8eb6fb-6c95-4452-9f98-963a04b0d64d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 797.957680] env[62600]: DEBUG nova.compute.manager [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 797.965792] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:b6:6b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e41070eb-3ac1-4ca9-a3d0-fd65893a97de', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ab8eb6fb-6c95-4452-9f98-963a04b0d64d', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 797.973050] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Creating folder: Project (03b37f47579d415384c4cd92bef5ca00). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 797.974510] env[62600]: DEBUG nova.compute.manager [req-821d556b-47f3-4ad8-9ae3-982ceb66c977 req-cce485c2-2134-436d-96d4-88bd83750240 service nova] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Received event network-changed-ab8eb6fb-6c95-4452-9f98-963a04b0d64d {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 797.974706] env[62600]: DEBUG nova.compute.manager [req-821d556b-47f3-4ad8-9ae3-982ceb66c977 req-cce485c2-2134-436d-96d4-88bd83750240 service nova] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Refreshing instance network info cache due to event network-changed-ab8eb6fb-6c95-4452-9f98-963a04b0d64d. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 797.974939] env[62600]: DEBUG oslo_concurrency.lockutils [req-821d556b-47f3-4ad8-9ae3-982ceb66c977 req-cce485c2-2134-436d-96d4-88bd83750240 service nova] Acquiring lock "refresh_cache-734e8a2b-faaa-4165-b0e9-69eeb3699b2d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.975098] env[62600]: DEBUG oslo_concurrency.lockutils [req-821d556b-47f3-4ad8-9ae3-982ceb66c977 req-cce485c2-2134-436d-96d4-88bd83750240 service nova] Acquired lock "refresh_cache-734e8a2b-faaa-4165-b0e9-69eeb3699b2d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.975270] env[62600]: DEBUG nova.network.neutron [req-821d556b-47f3-4ad8-9ae3-982ceb66c977 req-cce485c2-2134-436d-96d4-88bd83750240 service nova] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Refreshing network info cache for port ab8eb6fb-6c95-4452-9f98-963a04b0d64d {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 797.983321] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29cdd6d8-319f-4ef9-9e6c-fcd617c4ca30 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.991134] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4508d873-d189-4652-8557-6164fa204f3c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.012720] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': task-1222332, 'name': Rename_Task, 'duration_secs': 0.194411} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.012915] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222333, 'name': CreateVM_Task, 'duration_secs': 0.595054} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.013404] env[62600]: DEBUG oslo_vmware.api [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222334, 'name': ReconfigVM_Task, 'duration_secs': 0.430417} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.013615] env[62600]: DEBUG oslo_vmware.api [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Task: {'id': task-1222331, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.245306} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.015783] env[62600]: DEBUG nova.virt.hardware [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 798.015999] env[62600]: DEBUG nova.virt.hardware [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 798.016197] env[62600]: DEBUG nova.virt.hardware [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 798.016393] env[62600]: DEBUG nova.virt.hardware [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 798.016562] env[62600]: DEBUG nova.virt.hardware [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 798.016762] env[62600]: DEBUG nova.virt.hardware [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 798.017192] env[62600]: DEBUG nova.virt.hardware [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 798.017233] env[62600]: DEBUG nova.virt.hardware [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 798.017382] env[62600]: DEBUG nova.virt.hardware [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 798.017544] env[62600]: DEBUG nova.virt.hardware [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 798.017744] env[62600]: DEBUG nova.virt.hardware [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 798.018795] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 798.018986] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 798.019270] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Reconfigured VM instance instance-0000002b to detach disk 2000 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 798.019563] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 798.019772] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 798.019961] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 798.020145] env[62600]: INFO nova.compute.manager [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Took 3.30 seconds to destroy the instance on the hypervisor. [ 798.020811] env[62600]: DEBUG oslo.service.loopingcall [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.021077] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-330fcbee-2df2-40c4-8d1d-6e7d6fe36dd2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.027397] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-057f2221-c603-4ce4-9588-9b463c54351a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.027989] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.028205] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.028504] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 798.029300] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ffc58d5-2869-45f2-bd6c-21b13b24910f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.032476] env[62600]: DEBUG nova.compute.manager [-] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 798.032574] env[62600]: DEBUG nova.network.neutron [-] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 798.035623] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9dabcd6-1d75-4104-8de1-f68392374b6f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.039377] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20044532-a10a-46e4-bc8c-6e7ebde86ff7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.043260] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Created folder: Project (03b37f47579d415384c4cd92bef5ca00) in parent group-v264198. [ 798.043473] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Creating folder: Instances. Parent ref: group-v264243. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 798.052020] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e3c5dcf5-30e2-4e53-8233-7feeb028aa0e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.063242] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afff95a6-11d9-40d4-a93d-7e17d2031c3a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.074706] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] de4f9304-3357-4eaa-9c94-fe28bc554086/de4f9304-3357-4eaa-9c94-fe28bc554086.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 798.075058] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Waiting for the task: (returnval){ [ 798.075058] env[62600]: value = "task-1222336" [ 798.075058] env[62600]: _type = "Task" [ 798.075058] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.085791] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c01ae43-eca5-493b-8587-924a5751e3a4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.098932] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 798.098932] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520cc157-3d8f-1d30-914f-cfbc1f68f839" [ 798.098932] env[62600]: _type = "Task" [ 798.098932] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.099397] env[62600]: DEBUG nova.compute.provider_tree [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 798.105152] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Created folder: Instances in parent group-v264243. [ 798.105382] env[62600]: DEBUG oslo.service.loopingcall [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.118431] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 798.120027] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-56d04423-ddb4-41ce-b9d3-6583d4274095 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.136372] env[62600]: DEBUG oslo_vmware.api [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 798.136372] env[62600]: value = "task-1222338" [ 798.136372] env[62600]: _type = "Task" [ 798.136372] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.138961] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': task-1222336, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.150195] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520cc157-3d8f-1d30-914f-cfbc1f68f839, 'name': SearchDatastore_Task, 'duration_secs': 0.015192} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.152776] env[62600]: DEBUG nova.network.neutron [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Successfully updated port: 73800cd5-a9c3-41ae-b4fe-fbb4d32081b7 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 798.154503] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.154795] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 798.155124] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.155301] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.155533] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 798.155790] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 798.155790] env[62600]: value = "task-1222339" [ 798.155790] env[62600]: _type = "Task" [ 798.155790] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.159368] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1a2ea88d-75b4-4fec-b5d7-92a10f748815 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.163461] env[62600]: DEBUG oslo_vmware.api [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222338, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.174248] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222339, 'name': CreateVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.175675] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 798.175932] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 798.177023] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3368cf70-6561-40e0-a297-5bf571198636 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.182803] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 798.182803] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526023cf-514b-5952-0300-4ec2fbfb1a61" [ 798.182803] env[62600]: _type = "Task" [ 798.182803] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.192192] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526023cf-514b-5952-0300-4ec2fbfb1a61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.609976] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': task-1222336, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.615951] env[62600]: DEBUG nova.scheduler.client.report [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 798.628535] env[62600]: DEBUG nova.network.neutron [req-821d556b-47f3-4ad8-9ae3-982ceb66c977 req-cce485c2-2134-436d-96d4-88bd83750240 service nova] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Updated VIF entry in instance network info cache for port ab8eb6fb-6c95-4452-9f98-963a04b0d64d. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 798.628951] env[62600]: DEBUG nova.network.neutron [req-821d556b-47f3-4ad8-9ae3-982ceb66c977 req-cce485c2-2134-436d-96d4-88bd83750240 service nova] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Updating instance_info_cache with network_info: [{"id": "ab8eb6fb-6c95-4452-9f98-963a04b0d64d", "address": "fa:16:3e:81:b6:6b", "network": {"id": "2edc47d5-8cf8-4d69-8630-3fe0e39e1c06", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-386075803-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "03b37f47579d415384c4cd92bef5ca00", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e41070eb-3ac1-4ca9-a3d0-fd65893a97de", "external-id": "nsx-vlan-transportzone-596", "segmentation_id": 596, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapab8eb6fb-6c", "ovs_interfaceid": "ab8eb6fb-6c95-4452-9f98-963a04b0d64d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.654020] env[62600]: DEBUG oslo_vmware.api [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222338, 'name': ReconfigVM_Task, 'duration_secs': 0.43495} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.654020] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Reconfigured VM instance instance-0000002b to attach disk [datastore2] de4f9304-3357-4eaa-9c94-fe28bc554086/de4f9304-3357-4eaa-9c94-fe28bc554086.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 798.654020] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Updating instance 'de4f9304-3357-4eaa-9c94-fe28bc554086' progress to 50 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 798.659081] env[62600]: DEBUG oslo_concurrency.lockutils [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Acquiring lock "refresh_cache-58bf5b37-05dd-478f-8c0b-5475ab63bf71" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.659081] env[62600]: DEBUG oslo_concurrency.lockutils [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Acquired lock "refresh_cache-58bf5b37-05dd-478f-8c0b-5475ab63bf71" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.659081] env[62600]: DEBUG nova.network.neutron [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 798.671236] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222339, 'name': CreateVM_Task, 'duration_secs': 0.338932} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.675498] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 798.676378] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.676456] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.676822] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 798.677117] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55e2c087-1300-47dd-bb3a-890684440d33 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.684343] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Waiting for the task: (returnval){ [ 798.684343] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5235e823-bbd0-e168-5383-5ee409daa9e6" [ 798.684343] env[62600]: _type = "Task" [ 798.684343] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.693654] env[62600]: DEBUG oslo_vmware.rw_handles [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5216eced-7788-1fb4-a39e-094004af51a1/disk-0.vmdk. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 798.695309] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9a10dab-c4d5-4bb0-a0e4-3f307823bd69 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.704182] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526023cf-514b-5952-0300-4ec2fbfb1a61, 'name': SearchDatastore_Task, 'duration_secs': 0.01061} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.709482] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43a9bd69-6df7-4c95-be94-7a8783b40256 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.712102] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5235e823-bbd0-e168-5383-5ee409daa9e6, 'name': SearchDatastore_Task, 'duration_secs': 0.013786} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.713855] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.714190] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 798.714483] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.714774] env[62600]: DEBUG oslo_vmware.rw_handles [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5216eced-7788-1fb4-a39e-094004af51a1/disk-0.vmdk is in state: ready. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 798.715161] env[62600]: ERROR oslo_vmware.rw_handles [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5216eced-7788-1fb4-a39e-094004af51a1/disk-0.vmdk due to incomplete transfer. [ 798.716050] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-de8a5612-e0b5-4f24-b9ca-127c2e796741 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.718907] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 798.718907] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5205a19b-2e1c-1299-d198-f9bddb170f0e" [ 798.718907] env[62600]: _type = "Task" [ 798.718907] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.724660] env[62600]: DEBUG oslo_vmware.rw_handles [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5216eced-7788-1fb4-a39e-094004af51a1/disk-0.vmdk. {{(pid=62600) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 798.724972] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Uploaded image 6f9a3616-6c62-4597-af46-3d4f504e1193 to the Glance image server {{(pid=62600) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 798.727791] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Destroying the VM {{(pid=62600) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 798.731856] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-83b6f359-0344-444e-97fe-bbb33fd37618 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.732808] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5205a19b-2e1c-1299-d198-f9bddb170f0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.737972] env[62600]: DEBUG oslo_vmware.api [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 798.737972] env[62600]: value = "task-1222340" [ 798.737972] env[62600]: _type = "Task" [ 798.737972] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.746538] env[62600]: DEBUG oslo_vmware.api [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222340, 'name': Destroy_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.899040] env[62600]: DEBUG nova.network.neutron [-] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.109182] env[62600]: DEBUG oslo_vmware.api [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': task-1222336, 'name': PowerOnVM_Task, 'duration_secs': 0.577425} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.109479] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 799.109621] env[62600]: INFO nova.compute.manager [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Took 9.13 seconds to spawn the instance on the hypervisor. [ 799.109827] env[62600]: DEBUG nova.compute.manager [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 799.111120] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f1faea9-7834-4e27-a6bf-42dd356e6bde {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.120030] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.494s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.123761] env[62600]: DEBUG nova.compute.manager [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 799.123761] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.310s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.124530] env[62600]: INFO nova.compute.claims [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 799.131395] env[62600]: DEBUG oslo_concurrency.lockutils [req-821d556b-47f3-4ad8-9ae3-982ceb66c977 req-cce485c2-2134-436d-96d4-88bd83750240 service nova] Releasing lock "refresh_cache-734e8a2b-faaa-4165-b0e9-69eeb3699b2d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.159018] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b750145f-522c-478c-a285-6fbb3260418f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.184427] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aca7d48a-dffe-4a54-be69-8a1cb613554c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.204557] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Updating instance 'de4f9304-3357-4eaa-9c94-fe28bc554086' progress to 67 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 799.208490] env[62600]: DEBUG nova.network.neutron [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 799.230708] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5205a19b-2e1c-1299-d198-f9bddb170f0e, 'name': SearchDatastore_Task, 'duration_secs': 0.013613} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.232998] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.233357] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] d34c098d-a827-4dbd-96e4-e27d9d56b847/d34c098d-a827-4dbd-96e4-e27d9d56b847.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 799.233705] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.233952] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.234188] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-29ec58d3-c026-4c7b-bff3-d9e703155674 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.236291] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eb0f6254-dd03-4634-9c95-eef78f023bf6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.243438] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 799.243438] env[62600]: value = "task-1222341" [ 799.243438] env[62600]: _type = "Task" [ 799.243438] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.247019] env[62600]: DEBUG oslo_vmware.api [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222340, 'name': Destroy_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.250265] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.250420] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 799.251457] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c57e042-35c9-47b4-90c1-29a57a8bc79c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.256790] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222341, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.259622] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Waiting for the task: (returnval){ [ 799.259622] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52eee358-57f6-141d-160c-b9f53cff6481" [ 799.259622] env[62600]: _type = "Task" [ 799.259622] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.269111] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52eee358-57f6-141d-160c-b9f53cff6481, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.369843] env[62600]: DEBUG nova.network.neutron [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Updating instance_info_cache with network_info: [{"id": "73800cd5-a9c3-41ae-b4fe-fbb4d32081b7", "address": "fa:16:3e:59:30:00", "network": {"id": "fe613829-805d-4727-b0b1-d76610125c11", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1726586868-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cdf4ad7a2c214abd99b6e24550ac157e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad36dd36-1d2c-4f37-a259-98ef2e440794", "external-id": "nsx-vlan-transportzone-479", "segmentation_id": 479, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73800cd5-a9", "ovs_interfaceid": "73800cd5-a9c3-41ae-b4fe-fbb4d32081b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.401506] env[62600]: INFO nova.compute.manager [-] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Took 1.37 seconds to deallocate network for instance. [ 799.540377] env[62600]: DEBUG nova.compute.manager [req-cfafc1cb-c65c-4938-aba2-078ee71b7ffe req-9d1ba136-855f-4dd6-99d8-5bec2ae8f8a6 service nova] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Received event network-vif-plugged-73800cd5-a9c3-41ae-b4fe-fbb4d32081b7 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.540377] env[62600]: DEBUG oslo_concurrency.lockutils [req-cfafc1cb-c65c-4938-aba2-078ee71b7ffe req-9d1ba136-855f-4dd6-99d8-5bec2ae8f8a6 service nova] Acquiring lock "58bf5b37-05dd-478f-8c0b-5475ab63bf71-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.540547] env[62600]: DEBUG oslo_concurrency.lockutils [req-cfafc1cb-c65c-4938-aba2-078ee71b7ffe req-9d1ba136-855f-4dd6-99d8-5bec2ae8f8a6 service nova] Lock "58bf5b37-05dd-478f-8c0b-5475ab63bf71-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.540725] env[62600]: DEBUG oslo_concurrency.lockutils [req-cfafc1cb-c65c-4938-aba2-078ee71b7ffe req-9d1ba136-855f-4dd6-99d8-5bec2ae8f8a6 service nova] Lock "58bf5b37-05dd-478f-8c0b-5475ab63bf71-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.540883] env[62600]: DEBUG nova.compute.manager [req-cfafc1cb-c65c-4938-aba2-078ee71b7ffe req-9d1ba136-855f-4dd6-99d8-5bec2ae8f8a6 service nova] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] No waiting events found dispatching network-vif-plugged-73800cd5-a9c3-41ae-b4fe-fbb4d32081b7 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 799.541079] env[62600]: WARNING nova.compute.manager [req-cfafc1cb-c65c-4938-aba2-078ee71b7ffe req-9d1ba136-855f-4dd6-99d8-5bec2ae8f8a6 service nova] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Received unexpected event network-vif-plugged-73800cd5-a9c3-41ae-b4fe-fbb4d32081b7 for instance with vm_state building and task_state spawning. [ 799.541252] env[62600]: DEBUG nova.compute.manager [req-cfafc1cb-c65c-4938-aba2-078ee71b7ffe req-9d1ba136-855f-4dd6-99d8-5bec2ae8f8a6 service nova] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Received event network-changed-73800cd5-a9c3-41ae-b4fe-fbb4d32081b7 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.541410] env[62600]: DEBUG nova.compute.manager [req-cfafc1cb-c65c-4938-aba2-078ee71b7ffe req-9d1ba136-855f-4dd6-99d8-5bec2ae8f8a6 service nova] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Refreshing instance network info cache due to event network-changed-73800cd5-a9c3-41ae-b4fe-fbb4d32081b7. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 799.541577] env[62600]: DEBUG oslo_concurrency.lockutils [req-cfafc1cb-c65c-4938-aba2-078ee71b7ffe req-9d1ba136-855f-4dd6-99d8-5bec2ae8f8a6 service nova] Acquiring lock "refresh_cache-58bf5b37-05dd-478f-8c0b-5475ab63bf71" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.628407] env[62600]: DEBUG nova.compute.utils [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 799.634216] env[62600]: DEBUG nova.compute.manager [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 799.634390] env[62600]: DEBUG nova.network.neutron [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 799.636752] env[62600]: INFO nova.compute.manager [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Took 25.41 seconds to build instance. [ 799.716711] env[62600]: DEBUG nova.policy [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b24ebe18c1444edaa8000d83e5c4ff6d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '888577fb598f45c881757ce3ee52bfa3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 799.756966] env[62600]: DEBUG oslo_vmware.api [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222340, 'name': Destroy_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.765715] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222341, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.776031] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52eee358-57f6-141d-160c-b9f53cff6481, 'name': SearchDatastore_Task, 'duration_secs': 0.028431} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.777633] env[62600]: DEBUG nova.network.neutron [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Port fe02d314-372f-43fb-8d7c-cfc33c9e74bd binding to destination host cpu-1 is already ACTIVE {{(pid=62600) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 799.782019] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0932db26-643f-45d1-9930-287576de6d6a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.787798] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Waiting for the task: (returnval){ [ 799.787798] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b4dd6b-27ff-408d-eb8f-cf3ce757a7b3" [ 799.787798] env[62600]: _type = "Task" [ 799.787798] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.803690] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b4dd6b-27ff-408d-eb8f-cf3ce757a7b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.872785] env[62600]: DEBUG oslo_concurrency.lockutils [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Releasing lock "refresh_cache-58bf5b37-05dd-478f-8c0b-5475ab63bf71" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.873268] env[62600]: DEBUG nova.compute.manager [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Instance network_info: |[{"id": "73800cd5-a9c3-41ae-b4fe-fbb4d32081b7", "address": "fa:16:3e:59:30:00", "network": {"id": "fe613829-805d-4727-b0b1-d76610125c11", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1726586868-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cdf4ad7a2c214abd99b6e24550ac157e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad36dd36-1d2c-4f37-a259-98ef2e440794", "external-id": "nsx-vlan-transportzone-479", "segmentation_id": 479, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73800cd5-a9", "ovs_interfaceid": "73800cd5-a9c3-41ae-b4fe-fbb4d32081b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 799.873676] env[62600]: DEBUG oslo_concurrency.lockutils [req-cfafc1cb-c65c-4938-aba2-078ee71b7ffe req-9d1ba136-855f-4dd6-99d8-5bec2ae8f8a6 service nova] Acquired lock "refresh_cache-58bf5b37-05dd-478f-8c0b-5475ab63bf71" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.873945] env[62600]: DEBUG nova.network.neutron [req-cfafc1cb-c65c-4938-aba2-078ee71b7ffe req-9d1ba136-855f-4dd6-99d8-5bec2ae8f8a6 service nova] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Refreshing network info cache for port 73800cd5-a9c3-41ae-b4fe-fbb4d32081b7 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 799.875526] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:30:00', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad36dd36-1d2c-4f37-a259-98ef2e440794', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '73800cd5-a9c3-41ae-b4fe-fbb4d32081b7', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 799.884806] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Creating folder: Project (cdf4ad7a2c214abd99b6e24550ac157e). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 799.891828] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e794daa1-0724-46ba-ade9-d428bfcf9c00 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.903744] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Created folder: Project (cdf4ad7a2c214abd99b6e24550ac157e) in parent group-v264198. [ 799.904153] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Creating folder: Instances. Parent ref: group-v264246. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 799.904752] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4c81201e-84e2-41e0-8f22-c99cc85d8cf7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.908307] env[62600]: DEBUG oslo_concurrency.lockutils [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.915799] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Created folder: Instances in parent group-v264246. [ 799.916139] env[62600]: DEBUG oslo.service.loopingcall [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 799.916404] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 799.916710] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a33f3088-eca8-4d89-ba36-2460f82ac2b5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.942933] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 799.942933] env[62600]: value = "task-1222344" [ 799.942933] env[62600]: _type = "Task" [ 799.942933] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.956247] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222344, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.134740] env[62600]: DEBUG nova.compute.manager [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 800.140754] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e1942554-fd64-43b7-ab19-e4361ab31fcf tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Lock "92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.457s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.146190] env[62600]: DEBUG nova.network.neutron [req-cfafc1cb-c65c-4938-aba2-078ee71b7ffe req-9d1ba136-855f-4dd6-99d8-5bec2ae8f8a6 service nova] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Updated VIF entry in instance network info cache for port 73800cd5-a9c3-41ae-b4fe-fbb4d32081b7. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 800.146190] env[62600]: DEBUG nova.network.neutron [req-cfafc1cb-c65c-4938-aba2-078ee71b7ffe req-9d1ba136-855f-4dd6-99d8-5bec2ae8f8a6 service nova] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Updating instance_info_cache with network_info: [{"id": "73800cd5-a9c3-41ae-b4fe-fbb4d32081b7", "address": "fa:16:3e:59:30:00", "network": {"id": "fe613829-805d-4727-b0b1-d76610125c11", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1726586868-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cdf4ad7a2c214abd99b6e24550ac157e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad36dd36-1d2c-4f37-a259-98ef2e440794", "external-id": "nsx-vlan-transportzone-479", "segmentation_id": 479, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73800cd5-a9", "ovs_interfaceid": "73800cd5-a9c3-41ae-b4fe-fbb4d32081b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.260412] env[62600]: DEBUG oslo_vmware.api [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222340, 'name': Destroy_Task, 'duration_secs': 1.022861} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.261344] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Destroyed the VM [ 800.261771] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Deleting Snapshot of the VM instance {{(pid=62600) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 800.262142] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-8f9ae09d-1d30-472b-84b3-957ed52b3891 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.267855] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222341, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.797265} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.268463] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] d34c098d-a827-4dbd-96e4-e27d9d56b847/d34c098d-a827-4dbd-96e4-e27d9d56b847.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 800.268760] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 800.269083] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2fc3a172-9d28-40b9-9b62-79f81a02e627 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.275672] env[62600]: DEBUG oslo_vmware.api [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 800.275672] env[62600]: value = "task-1222345" [ 800.275672] env[62600]: _type = "Task" [ 800.275672] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.287557] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 800.287557] env[62600]: value = "task-1222346" [ 800.287557] env[62600]: _type = "Task" [ 800.287557] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.302141] env[62600]: DEBUG oslo_vmware.api [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222345, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.307559] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222346, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.314503] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b4dd6b-27ff-408d-eb8f-cf3ce757a7b3, 'name': SearchDatastore_Task, 'duration_secs': 0.057063} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.315064] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.315471] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 734e8a2b-faaa-4165-b0e9-69eeb3699b2d/734e8a2b-faaa-4165-b0e9-69eeb3699b2d.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 800.315856] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bf44a0f6-e834-4ca9-b67e-b8d863714863 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.321920] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Waiting for the task: (returnval){ [ 800.321920] env[62600]: value = "task-1222347" [ 800.321920] env[62600]: _type = "Task" [ 800.321920] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.346357] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': task-1222347, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.452315] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222344, 'name': CreateVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.645424] env[62600]: DEBUG nova.compute.manager [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 800.649671] env[62600]: DEBUG oslo_concurrency.lockutils [req-cfafc1cb-c65c-4938-aba2-078ee71b7ffe req-9d1ba136-855f-4dd6-99d8-5bec2ae8f8a6 service nova] Releasing lock "refresh_cache-58bf5b37-05dd-478f-8c0b-5475ab63bf71" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.652432] env[62600]: DEBUG nova.compute.manager [req-cfafc1cb-c65c-4938-aba2-078ee71b7ffe req-9d1ba136-855f-4dd6-99d8-5bec2ae8f8a6 service nova] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Received event network-vif-deleted-6c00cb70-1e44-47a1-a1da-6273976cda79 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 800.652432] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c90758c9-ef54-4ccf-a5e1-2d4b9cf53efe {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.659561] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1aa8bfd-fcae-4586-9fc6-c95e210dd025 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.701897] env[62600]: DEBUG nova.network.neutron [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Successfully created port: b4b6687a-b518-40c1-99a5-7d1ffe49e1f2 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 800.704425] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f00d97b-16fe-4feb-9d29-bc5fb1e64de8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.715258] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b86ab5e0-f84e-49f5-a25d-00ab2aaaec7f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.729952] env[62600]: DEBUG nova.compute.provider_tree [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.786809] env[62600]: DEBUG oslo_vmware.api [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222345, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.803980] env[62600]: DEBUG oslo_concurrency.lockutils [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "de4f9304-3357-4eaa-9c94-fe28bc554086-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.804276] env[62600]: DEBUG oslo_concurrency.lockutils [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "de4f9304-3357-4eaa-9c94-fe28bc554086-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.804504] env[62600]: DEBUG oslo_concurrency.lockutils [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "de4f9304-3357-4eaa-9c94-fe28bc554086-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.821801] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222346, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068691} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.822729] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 800.823702] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735f1441-3431-4d57-bf07-fd2fcd59e5a4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.852231] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] d34c098d-a827-4dbd-96e4-e27d9d56b847/d34c098d-a827-4dbd-96e4-e27d9d56b847.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 800.852602] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': task-1222347, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.853224] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9c68798-ec86-40c9-85dd-8c849be99dee {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.873087] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 800.873087] env[62600]: value = "task-1222348" [ 800.873087] env[62600]: _type = "Task" [ 800.873087] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.881911] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222348, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.952791] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222344, 'name': CreateVM_Task, 'duration_secs': 0.540101} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.953015] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 800.953760] env[62600]: DEBUG oslo_concurrency.lockutils [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.953974] env[62600]: DEBUG oslo_concurrency.lockutils [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.954328] env[62600]: DEBUG oslo_concurrency.lockutils [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 800.954592] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5af3d7b9-757a-4144-ad13-64878755ab7b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.958988] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Waiting for the task: (returnval){ [ 800.958988] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5238fe1e-dd48-02a8-addc-c62d4d556425" [ 800.958988] env[62600]: _type = "Task" [ 800.958988] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.968267] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5238fe1e-dd48-02a8-addc-c62d4d556425, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.152922] env[62600]: DEBUG nova.compute.manager [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 801.187617] env[62600]: DEBUG nova.virt.hardware [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 801.187915] env[62600]: DEBUG nova.virt.hardware [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 801.188225] env[62600]: DEBUG nova.virt.hardware [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 801.188476] env[62600]: DEBUG nova.virt.hardware [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 801.188668] env[62600]: DEBUG nova.virt.hardware [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 801.188862] env[62600]: DEBUG nova.virt.hardware [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 801.189179] env[62600]: DEBUG nova.virt.hardware [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 801.189399] env[62600]: DEBUG nova.virt.hardware [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 801.190942] env[62600]: DEBUG nova.virt.hardware [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 801.190942] env[62600]: DEBUG nova.virt.hardware [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 801.190942] env[62600]: DEBUG nova.virt.hardware [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 801.191071] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bbc8507-4a8f-473e-80f5-b8113414e919 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.201449] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.211099] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5834bcb-8fbb-4467-bde5-36257470e953 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.233405] env[62600]: DEBUG nova.scheduler.client.report [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 801.286795] env[62600]: DEBUG oslo_vmware.api [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222345, 'name': RemoveSnapshot_Task, 'duration_secs': 0.845373} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.288437] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Deleted Snapshot of the VM instance {{(pid=62600) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 801.288437] env[62600]: INFO nova.compute.manager [None req-83c7651f-77d8-488b-8940-5909a7085ca1 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Took 17.23 seconds to snapshot the instance on the hypervisor. [ 801.338468] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': task-1222347, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.64333} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.338815] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 734e8a2b-faaa-4165-b0e9-69eeb3699b2d/734e8a2b-faaa-4165-b0e9-69eeb3699b2d.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 801.339079] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 801.339348] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-710ce37a-cdc3-44e7-9a45-6ee15f90f638 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.347827] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Waiting for the task: (returnval){ [ 801.347827] env[62600]: value = "task-1222349" [ 801.347827] env[62600]: _type = "Task" [ 801.347827] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.356095] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': task-1222349, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.387816] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222348, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.468812] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5238fe1e-dd48-02a8-addc-c62d4d556425, 'name': SearchDatastore_Task, 'duration_secs': 0.024681} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.469171] env[62600]: DEBUG oslo_concurrency.lockutils [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.469413] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 801.469652] env[62600]: DEBUG oslo_concurrency.lockutils [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.469833] env[62600]: DEBUG oslo_concurrency.lockutils [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.470036] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 801.470315] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0fe7d7c8-fefb-4090-85de-c460fd824ecc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.478463] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 801.478649] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 801.479394] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54946ba8-97c0-4e75-9fbd-98a338166a81 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.484546] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Waiting for the task: (returnval){ [ 801.484546] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]522c96bb-c680-8f72-4cf2-25754bec1950" [ 801.484546] env[62600]: _type = "Task" [ 801.484546] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.491880] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]522c96bb-c680-8f72-4cf2-25754bec1950, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.505528] env[62600]: DEBUG nova.compute.manager [req-fe219ed4-b7cd-4477-8856-cbead901e2a0 req-1b10c445-9297-49b4-9732-fa8101fd6406 service nova] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Received event network-changed-cb1945e8-09df-4f1f-ba94-77e84b24f494 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 801.505749] env[62600]: DEBUG nova.compute.manager [req-fe219ed4-b7cd-4477-8856-cbead901e2a0 req-1b10c445-9297-49b4-9732-fa8101fd6406 service nova] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Refreshing instance network info cache due to event network-changed-cb1945e8-09df-4f1f-ba94-77e84b24f494. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 801.505931] env[62600]: DEBUG oslo_concurrency.lockutils [req-fe219ed4-b7cd-4477-8856-cbead901e2a0 req-1b10c445-9297-49b4-9732-fa8101fd6406 service nova] Acquiring lock "refresh_cache-92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.506084] env[62600]: DEBUG oslo_concurrency.lockutils [req-fe219ed4-b7cd-4477-8856-cbead901e2a0 req-1b10c445-9297-49b4-9732-fa8101fd6406 service nova] Acquired lock "refresh_cache-92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.506264] env[62600]: DEBUG nova.network.neutron [req-fe219ed4-b7cd-4477-8856-cbead901e2a0 req-1b10c445-9297-49b4-9732-fa8101fd6406 service nova] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Refreshing network info cache for port cb1945e8-09df-4f1f-ba94-77e84b24f494 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 801.741721] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.618s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.742208] env[62600]: DEBUG nova.compute.manager [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 801.745354] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.730s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.857811] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': task-1222349, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087717} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.858148] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 801.858985] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7da29d-ebc5-4c95-9e2f-5193b74f8d1e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.884532] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 734e8a2b-faaa-4165-b0e9-69eeb3699b2d/734e8a2b-faaa-4165-b0e9-69eeb3699b2d.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 801.885318] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a212d0c2-25e9-4ef7-a98d-60c7e02a142a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.908772] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222348, 'name': ReconfigVM_Task, 'duration_secs': 0.570175} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.910437] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Reconfigured VM instance instance-00000033 to attach disk [datastore1] d34c098d-a827-4dbd-96e4-e27d9d56b847/d34c098d-a827-4dbd-96e4-e27d9d56b847.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 801.911273] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Waiting for the task: (returnval){ [ 801.911273] env[62600]: value = "task-1222350" [ 801.911273] env[62600]: _type = "Task" [ 801.911273] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.911630] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e726224f-2fab-44c4-bc87-55aeab706c9d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.925222] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': task-1222350, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.926647] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 801.926647] env[62600]: value = "task-1222351" [ 801.926647] env[62600]: _type = "Task" [ 801.926647] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.935883] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222351, 'name': Rename_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.994825] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]522c96bb-c680-8f72-4cf2-25754bec1950, 'name': SearchDatastore_Task, 'duration_secs': 0.013684} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.995690] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-210cc7d4-4c46-42a9-b9f1-db3a81a641f5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.000964] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Waiting for the task: (returnval){ [ 802.000964] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e5cc78-f125-bdb6-d8cd-d3a3cdbe4e75" [ 802.000964] env[62600]: _type = "Task" [ 802.000964] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.011068] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e5cc78-f125-bdb6-d8cd-d3a3cdbe4e75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.075680] env[62600]: DEBUG oslo_concurrency.lockutils [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "refresh_cache-de4f9304-3357-4eaa-9c94-fe28bc554086" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.075979] env[62600]: DEBUG oslo_concurrency.lockutils [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquired lock "refresh_cache-de4f9304-3357-4eaa-9c94-fe28bc554086" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.076215] env[62600]: DEBUG nova.network.neutron [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 802.226768] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6029d55c-54d5-4fc0-bdf2-165be829f86c tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquiring lock "92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.227885] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6029d55c-54d5-4fc0-bdf2-165be829f86c tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Lock "92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.227885] env[62600]: INFO nova.compute.manager [None req-6029d55c-54d5-4fc0-bdf2-165be829f86c tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Rebooting instance [ 802.247445] env[62600]: DEBUG nova.compute.utils [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 802.248775] env[62600]: DEBUG nova.compute.manager [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 802.248960] env[62600]: DEBUG nova.network.neutron [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 802.319519] env[62600]: DEBUG nova.network.neutron [req-fe219ed4-b7cd-4477-8856-cbead901e2a0 req-1b10c445-9297-49b4-9732-fa8101fd6406 service nova] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Updated VIF entry in instance network info cache for port cb1945e8-09df-4f1f-ba94-77e84b24f494. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 802.319909] env[62600]: DEBUG nova.network.neutron [req-fe219ed4-b7cd-4477-8856-cbead901e2a0 req-1b10c445-9297-49b4-9732-fa8101fd6406 service nova] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Updating instance_info_cache with network_info: [{"id": "cb1945e8-09df-4f1f-ba94-77e84b24f494", "address": "fa:16:3e:63:98:19", "network": {"id": "f9decfe0-071c-4ff4-b65a-ea34be9fc487", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-439991991-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "583e67fe6a444511946ce9fac5c66420", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1ce8361b-fd8e-4971-a37f-b84a4f77db19", "external-id": "nsx-vlan-transportzone-255", "segmentation_id": 255, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb1945e8-09", "ovs_interfaceid": "cb1945e8-09df-4f1f-ba94-77e84b24f494", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.327150] env[62600]: DEBUG nova.policy [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1acacdc5f3a94ac7a7c446d3fc915d38', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7a393dbb43124b0f824ffea7665ab55b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 802.425346] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': task-1222350, 'name': ReconfigVM_Task, 'duration_secs': 0.295598} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.425346] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 734e8a2b-faaa-4165-b0e9-69eeb3699b2d/734e8a2b-faaa-4165-b0e9-69eeb3699b2d.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 802.425346] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7a6295bc-1b7c-47be-a08e-a1b0c6f47aa0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.437611] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222351, 'name': Rename_Task, 'duration_secs': 0.180274} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.439081] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 802.439506] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Waiting for the task: (returnval){ [ 802.439506] env[62600]: value = "task-1222352" [ 802.439506] env[62600]: _type = "Task" [ 802.439506] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.443023] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6174e1dc-0112-4229-a944-49cd42e3e92d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.455050] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': task-1222352, 'name': Rename_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.458500] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 802.458500] env[62600]: value = "task-1222353" [ 802.458500] env[62600]: _type = "Task" [ 802.458500] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.467355] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222353, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.515932] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e5cc78-f125-bdb6-d8cd-d3a3cdbe4e75, 'name': SearchDatastore_Task, 'duration_secs': 0.02716} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.516261] env[62600]: DEBUG oslo_concurrency.lockutils [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.516526] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 58bf5b37-05dd-478f-8c0b-5475ab63bf71/58bf5b37-05dd-478f-8c0b-5475ab63bf71.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 802.516791] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-28a430ad-5085-41a0-8c86-df6b7da7ab38 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.523154] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Waiting for the task: (returnval){ [ 802.523154] env[62600]: value = "task-1222354" [ 802.523154] env[62600]: _type = "Task" [ 802.523154] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.531180] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': task-1222354, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.592044] env[62600]: DEBUG nova.network.neutron [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Successfully created port: 1542d23a-48d6-479e-af46-e528ef96651f {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 802.738172] env[62600]: DEBUG nova.network.neutron [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Successfully updated port: b4b6687a-b518-40c1-99a5-7d1ffe49e1f2 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 802.754290] env[62600]: DEBUG nova.compute.manager [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 802.764613] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Applying migration context for instance de4f9304-3357-4eaa-9c94-fe28bc554086 as it has an incoming, in-progress migration a871b6f5-809b-4c75-97f7-3f72b2a3dc6a. Migration status is post-migrating {{(pid=62600) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 802.767117] env[62600]: INFO nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Updating resource usage from migration a871b6f5-809b-4c75-97f7-3f72b2a3dc6a [ 802.771354] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6029d55c-54d5-4fc0-bdf2-165be829f86c tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquiring lock "refresh_cache-92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.806235] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 92cf1acb-b471-4add-a73f-7e9e94fbaaa4 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 802.806235] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance efff4d72-4673-4a8d-9a81-be9d700ff881 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 802.806235] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance a358a3cb-deda-419a-aa3c-ce7aeb534240 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 802.806235] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance e25c631d-3d1b-40d2-9fb1-a65431f991a9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 802.806436] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 802.806436] env[62600]: WARNING nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance dcbe8e2d-6ae9-465f-8394-3978ee61a15b is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 802.806436] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 802.806436] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance d34c098d-a827-4dbd-96e4-e27d9d56b847 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 802.806550] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 802.806982] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 734e8a2b-faaa-4165-b0e9-69eeb3699b2d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 802.807310] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 58bf5b37-05dd-478f-8c0b-5475ab63bf71 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 802.807488] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance afe838cc-e086-4986-87ec-4e1266bcaf60 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 802.807738] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 802.828317] env[62600]: DEBUG oslo_concurrency.lockutils [req-fe219ed4-b7cd-4477-8856-cbead901e2a0 req-1b10c445-9297-49b4-9732-fa8101fd6406 service nova] Releasing lock "refresh_cache-92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.828769] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6029d55c-54d5-4fc0-bdf2-165be829f86c tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquired lock "refresh_cache-92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.828985] env[62600]: DEBUG nova.network.neutron [None req-6029d55c-54d5-4fc0-bdf2-165be829f86c tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 802.954595] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': task-1222352, 'name': Rename_Task, 'duration_secs': 0.179417} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.960507] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 802.960507] env[62600]: DEBUG nova.network.neutron [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Updating instance_info_cache with network_info: [{"id": "fe02d314-372f-43fb-8d7c-cfc33c9e74bd", "address": "fa:16:3e:86:22:ec", "network": {"id": "d946b94e-9827-4c22-af05-072fcaa3d0e6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.235", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ec39f2547f624ed0b240d084b949694a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe02d314-37", "ovs_interfaceid": "fe02d314-372f-43fb-8d7c-cfc33c9e74bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.960787] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-276d693b-8718-427c-b940-3fe9a7a11a86 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.969919] env[62600]: DEBUG oslo_vmware.api [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222353, 'name': PowerOnVM_Task, 'duration_secs': 0.484263} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.973089] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 802.973505] env[62600]: DEBUG nova.compute.manager [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 802.974425] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Waiting for the task: (returnval){ [ 802.974425] env[62600]: value = "task-1222355" [ 802.974425] env[62600]: _type = "Task" [ 802.974425] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.975460] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc9b13b-fec2-4f4f-83aa-c0c9d5f65af6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.999678] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': task-1222355, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.033035] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': task-1222354, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.242187] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquiring lock "refresh_cache-afe838cc-e086-4986-87ec-4e1266bcaf60" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.242187] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquired lock "refresh_cache-afe838cc-e086-4986-87ec-4e1266bcaf60" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.242187] env[62600]: DEBUG nova.network.neutron [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 803.310910] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance bd9b7351-01a2-429d-a860-d9ff6855eefc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 803.461035] env[62600]: DEBUG oslo_concurrency.lockutils [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Releasing lock "refresh_cache-de4f9304-3357-4eaa-9c94-fe28bc554086" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.495504] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': task-1222355, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.507289] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.533060] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': task-1222354, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.574206} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.533492] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 58bf5b37-05dd-478f-8c0b-5475ab63bf71/58bf5b37-05dd-478f-8c0b-5475ab63bf71.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 803.533728] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 803.533983] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5026ebce-3195-455b-aa53-f53f042a6e56 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.540016] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Waiting for the task: (returnval){ [ 803.540016] env[62600]: value = "task-1222356" [ 803.540016] env[62600]: _type = "Task" [ 803.540016] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.552760] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': task-1222356, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.558976] env[62600]: DEBUG nova.network.neutron [None req-6029d55c-54d5-4fc0-bdf2-165be829f86c tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Updating instance_info_cache with network_info: [{"id": "cb1945e8-09df-4f1f-ba94-77e84b24f494", "address": "fa:16:3e:63:98:19", "network": {"id": "f9decfe0-071c-4ff4-b65a-ea34be9fc487", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-439991991-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "583e67fe6a444511946ce9fac5c66420", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1ce8361b-fd8e-4971-a37f-b84a4f77db19", "external-id": "nsx-vlan-transportzone-255", "segmentation_id": 255, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb1945e8-09", "ovs_interfaceid": "cb1945e8-09df-4f1f-ba94-77e84b24f494", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.679020] env[62600]: DEBUG nova.compute.manager [req-22dcd0ca-97b2-41e8-ad81-88cbd993e9cd req-2c016f20-1786-4eac-b8d6-d617ed5692cb service nova] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Received event network-vif-plugged-b4b6687a-b518-40c1-99a5-7d1ffe49e1f2 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 803.679020] env[62600]: DEBUG oslo_concurrency.lockutils [req-22dcd0ca-97b2-41e8-ad81-88cbd993e9cd req-2c016f20-1786-4eac-b8d6-d617ed5692cb service nova] Acquiring lock "afe838cc-e086-4986-87ec-4e1266bcaf60-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.679020] env[62600]: DEBUG oslo_concurrency.lockutils [req-22dcd0ca-97b2-41e8-ad81-88cbd993e9cd req-2c016f20-1786-4eac-b8d6-d617ed5692cb service nova] Lock "afe838cc-e086-4986-87ec-4e1266bcaf60-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.679020] env[62600]: DEBUG oslo_concurrency.lockutils [req-22dcd0ca-97b2-41e8-ad81-88cbd993e9cd req-2c016f20-1786-4eac-b8d6-d617ed5692cb service nova] Lock "afe838cc-e086-4986-87ec-4e1266bcaf60-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.679020] env[62600]: DEBUG nova.compute.manager [req-22dcd0ca-97b2-41e8-ad81-88cbd993e9cd req-2c016f20-1786-4eac-b8d6-d617ed5692cb service nova] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] No waiting events found dispatching network-vif-plugged-b4b6687a-b518-40c1-99a5-7d1ffe49e1f2 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 803.679497] env[62600]: WARNING nova.compute.manager [req-22dcd0ca-97b2-41e8-ad81-88cbd993e9cd req-2c016f20-1786-4eac-b8d6-d617ed5692cb service nova] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Received unexpected event network-vif-plugged-b4b6687a-b518-40c1-99a5-7d1ffe49e1f2 for instance with vm_state building and task_state spawning. [ 803.679497] env[62600]: DEBUG nova.compute.manager [req-22dcd0ca-97b2-41e8-ad81-88cbd993e9cd req-2c016f20-1786-4eac-b8d6-d617ed5692cb service nova] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Received event network-changed-b4b6687a-b518-40c1-99a5-7d1ffe49e1f2 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 803.679497] env[62600]: DEBUG nova.compute.manager [req-22dcd0ca-97b2-41e8-ad81-88cbd993e9cd req-2c016f20-1786-4eac-b8d6-d617ed5692cb service nova] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Refreshing instance network info cache due to event network-changed-b4b6687a-b518-40c1-99a5-7d1ffe49e1f2. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 803.679497] env[62600]: DEBUG oslo_concurrency.lockutils [req-22dcd0ca-97b2-41e8-ad81-88cbd993e9cd req-2c016f20-1786-4eac-b8d6-d617ed5692cb service nova] Acquiring lock "refresh_cache-afe838cc-e086-4986-87ec-4e1266bcaf60" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.763717] env[62600]: DEBUG nova.compute.manager [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 803.775771] env[62600]: DEBUG nova.network.neutron [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 803.791320] env[62600]: DEBUG nova.virt.hardware [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 803.791576] env[62600]: DEBUG nova.virt.hardware [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 803.791742] env[62600]: DEBUG nova.virt.hardware [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 803.792550] env[62600]: DEBUG nova.virt.hardware [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 803.792550] env[62600]: DEBUG nova.virt.hardware [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 803.792550] env[62600]: DEBUG nova.virt.hardware [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 803.792550] env[62600]: DEBUG nova.virt.hardware [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 803.792691] env[62600]: DEBUG nova.virt.hardware [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 803.792767] env[62600]: DEBUG nova.virt.hardware [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 803.792919] env[62600]: DEBUG nova.virt.hardware [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 803.793121] env[62600]: DEBUG nova.virt.hardware [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 803.794187] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b5c5d24-cfeb-4ac3-b873-997d67feda49 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.804236] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38ead890-e729-44dd-812f-a3c0aa86e2c0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.817648] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 4e1a376f-6619-4c35-b75f-b45db75815cc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 803.921914] env[62600]: DEBUG nova.network.neutron [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Updating instance_info_cache with network_info: [{"id": "b4b6687a-b518-40c1-99a5-7d1ffe49e1f2", "address": "fa:16:3e:2a:79:1c", "network": {"id": "102eb76e-a1ff-4b27-b285-516053dcad35", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-543226554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "888577fb598f45c881757ce3ee52bfa3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4b6687a-b5", "ovs_interfaceid": "b4b6687a-b518-40c1-99a5-7d1ffe49e1f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.986903] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a88b40c3-1888-4811-aff6-dcc560a16e4b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.992194] env[62600]: DEBUG oslo_vmware.api [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': task-1222355, 'name': PowerOnVM_Task, 'duration_secs': 0.530225} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.992842] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 803.993063] env[62600]: INFO nova.compute.manager [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Took 8.85 seconds to spawn the instance on the hypervisor. [ 803.993250] env[62600]: DEBUG nova.compute.manager [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 803.993949] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76363305-1c45-480a-a032-105ce91ca90f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.010241] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c22e687-2aad-4091-b9c8-b5e6c0046f22 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.020851] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Updating instance 'de4f9304-3357-4eaa-9c94-fe28bc554086' progress to 83 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 804.049097] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': task-1222356, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.19006} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.049954] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 804.050759] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cc6fb84-59e9-4aa5-8668-4a8ac3e0c808 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.065270] env[62600]: DEBUG nova.compute.manager [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 804.065727] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6029d55c-54d5-4fc0-bdf2-165be829f86c tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Releasing lock "refresh_cache-92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.075468] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 58bf5b37-05dd-478f-8c0b-5475ab63bf71/58bf5b37-05dd-478f-8c0b-5475ab63bf71.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 804.075950] env[62600]: DEBUG nova.compute.manager [None req-6029d55c-54d5-4fc0-bdf2-165be829f86c tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 804.076919] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99de689d-4111-496d-a0fc-46b62a184cc3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.080042] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-afadd066-d142-4931-9697-2e5069f90422 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.094657] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b242aed4-8917-41e9-8bd0-99c4550e03f3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.104987] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Waiting for the task: (returnval){ [ 804.104987] env[62600]: value = "task-1222357" [ 804.104987] env[62600]: _type = "Task" [ 804.104987] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.115732] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': task-1222357, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.233387] env[62600]: DEBUG nova.compute.manager [req-057ac512-2541-411f-8a15-271397d8c234 req-8018d457-de7f-46dc-8d5a-9686f6727e9d service nova] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Received event network-vif-plugged-1542d23a-48d6-479e-af46-e528ef96651f {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.233617] env[62600]: DEBUG oslo_concurrency.lockutils [req-057ac512-2541-411f-8a15-271397d8c234 req-8018d457-de7f-46dc-8d5a-9686f6727e9d service nova] Acquiring lock "8f3e87a5-bb70-4e50-9fea-6bf53946b8dd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.233841] env[62600]: DEBUG oslo_concurrency.lockutils [req-057ac512-2541-411f-8a15-271397d8c234 req-8018d457-de7f-46dc-8d5a-9686f6727e9d service nova] Lock "8f3e87a5-bb70-4e50-9fea-6bf53946b8dd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.234067] env[62600]: DEBUG oslo_concurrency.lockutils [req-057ac512-2541-411f-8a15-271397d8c234 req-8018d457-de7f-46dc-8d5a-9686f6727e9d service nova] Lock "8f3e87a5-bb70-4e50-9fea-6bf53946b8dd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.234522] env[62600]: DEBUG nova.compute.manager [req-057ac512-2541-411f-8a15-271397d8c234 req-8018d457-de7f-46dc-8d5a-9686f6727e9d service nova] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] No waiting events found dispatching network-vif-plugged-1542d23a-48d6-479e-af46-e528ef96651f {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 804.234714] env[62600]: WARNING nova.compute.manager [req-057ac512-2541-411f-8a15-271397d8c234 req-8018d457-de7f-46dc-8d5a-9686f6727e9d service nova] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Received unexpected event network-vif-plugged-1542d23a-48d6-479e-af46-e528ef96651f for instance with vm_state building and task_state spawning. [ 804.267463] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquiring lock "d34c098d-a827-4dbd-96e4-e27d9d56b847" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.270048] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "d34c098d-a827-4dbd-96e4-e27d9d56b847" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.270048] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquiring lock "d34c098d-a827-4dbd-96e4-e27d9d56b847-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.270048] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "d34c098d-a827-4dbd-96e4-e27d9d56b847-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.270048] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "d34c098d-a827-4dbd-96e4-e27d9d56b847-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.272838] env[62600]: INFO nova.compute.manager [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Terminating instance [ 804.276991] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquiring lock "refresh_cache-d34c098d-a827-4dbd-96e4-e27d9d56b847" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.277361] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquired lock "refresh_cache-d34c098d-a827-4dbd-96e4-e27d9d56b847" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.277726] env[62600]: DEBUG nova.network.neutron [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 804.320885] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance dde6df10-618a-40a8-b33f-efc0ca3a9287 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 804.379399] env[62600]: DEBUG nova.network.neutron [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Successfully updated port: 1542d23a-48d6-479e-af46-e528ef96651f {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 804.424198] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Releasing lock "refresh_cache-afe838cc-e086-4986-87ec-4e1266bcaf60" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.424753] env[62600]: DEBUG nova.compute.manager [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Instance network_info: |[{"id": "b4b6687a-b518-40c1-99a5-7d1ffe49e1f2", "address": "fa:16:3e:2a:79:1c", "network": {"id": "102eb76e-a1ff-4b27-b285-516053dcad35", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-543226554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "888577fb598f45c881757ce3ee52bfa3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4b6687a-b5", "ovs_interfaceid": "b4b6687a-b518-40c1-99a5-7d1ffe49e1f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 804.425138] env[62600]: DEBUG oslo_concurrency.lockutils [req-22dcd0ca-97b2-41e8-ad81-88cbd993e9cd req-2c016f20-1786-4eac-b8d6-d617ed5692cb service nova] Acquired lock "refresh_cache-afe838cc-e086-4986-87ec-4e1266bcaf60" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.425368] env[62600]: DEBUG nova.network.neutron [req-22dcd0ca-97b2-41e8-ad81-88cbd993e9cd req-2c016f20-1786-4eac-b8d6-d617ed5692cb service nova] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Refreshing network info cache for port b4b6687a-b518-40c1-99a5-7d1ffe49e1f2 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 804.429076] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2a:79:1c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3a0ddd7d-c321-4187-bdd8-b19044ea2c4a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b4b6687a-b518-40c1-99a5-7d1ffe49e1f2', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 804.437018] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Creating folder: Project (888577fb598f45c881757ce3ee52bfa3). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 804.437190] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-45bc5ae7-a8c9-4203-8953-7c553567b1c4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.449069] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Created folder: Project (888577fb598f45c881757ce3ee52bfa3) in parent group-v264198. [ 804.449293] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Creating folder: Instances. Parent ref: group-v264249. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 804.449531] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-738a55bf-40c3-4d6f-a417-3e6dd1e6a084 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.457560] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Created folder: Instances in parent group-v264249. [ 804.457804] env[62600]: DEBUG oslo.service.loopingcall [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 804.458403] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 804.458403] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4b5e9882-3ab4-44e6-9874-93526aad0595 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.480968] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 804.480968] env[62600]: value = "task-1222360" [ 804.480968] env[62600]: _type = "Task" [ 804.480968] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.488847] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222360, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.531777] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 804.533049] env[62600]: INFO nova.compute.manager [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Took 26.87 seconds to build instance. [ 804.533318] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ac381f44-6383-4b06-910a-411bea0c3de2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.539981] env[62600]: DEBUG oslo_vmware.api [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 804.539981] env[62600]: value = "task-1222361" [ 804.539981] env[62600]: _type = "Task" [ 804.539981] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.555331] env[62600]: DEBUG oslo_vmware.api [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222361, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.608863] env[62600]: INFO nova.compute.manager [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] instance snapshotting [ 804.617849] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed9aa03-3d28-4315-b929-30921b68f826 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.625646] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': task-1222357, 'name': ReconfigVM_Task, 'duration_secs': 0.325952} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.638921] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 58bf5b37-05dd-478f-8c0b-5475ab63bf71/58bf5b37-05dd-478f-8c0b-5475ab63bf71.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 804.639738] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1289bd43-462c-4480-a287-9275da12c85b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.642073] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c7ef68b-4cdf-4e1b-8048-cf5cf923de5a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.653790] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Waiting for the task: (returnval){ [ 804.653790] env[62600]: value = "task-1222362" [ 804.653790] env[62600]: _type = "Task" [ 804.653790] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.665274] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': task-1222362, 'name': Rename_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.799490] env[62600]: DEBUG nova.network.neutron [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 804.825687] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance fb6f8e36-8d24-45ea-a6e4-4d768c3b232d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 804.857679] env[62600]: DEBUG nova.network.neutron [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.882921] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Acquiring lock "refresh_cache-8f3e87a5-bb70-4e50-9fea-6bf53946b8dd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.883087] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Acquired lock "refresh_cache-8f3e87a5-bb70-4e50-9fea-6bf53946b8dd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.883336] env[62600]: DEBUG nova.network.neutron [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 804.978342] env[62600]: DEBUG oslo_concurrency.lockutils [None req-aa44dedc-e4b2-43ad-8fca-61e0bc902df2 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Acquiring lock "interface-734e8a2b-faaa-4165-b0e9-69eeb3699b2d-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.978790] env[62600]: DEBUG oslo_concurrency.lockutils [None req-aa44dedc-e4b2-43ad-8fca-61e0bc902df2 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Lock "interface-734e8a2b-faaa-4165-b0e9-69eeb3699b2d-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.979301] env[62600]: DEBUG nova.objects.instance [None req-aa44dedc-e4b2-43ad-8fca-61e0bc902df2 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Lazy-loading 'flavor' on Instance uuid 734e8a2b-faaa-4165-b0e9-69eeb3699b2d {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 804.993229] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222360, 'name': CreateVM_Task, 'duration_secs': 0.279099} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.993414] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 804.994172] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.994352] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.994708] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 804.994982] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2792df8a-2a13-4b09-a9ce-c06b9622e2e4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.001524] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 805.001524] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b952ae-b0fd-1cc0-d20b-527dddac91ce" [ 805.001524] env[62600]: _type = "Task" [ 805.001524] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.014349] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b952ae-b0fd-1cc0-d20b-527dddac91ce, 'name': SearchDatastore_Task, 'duration_secs': 0.010607} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.015187] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.015595] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 805.015671] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.015825] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.016037] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 805.016604] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f0c4aafb-fee4-49a2-b272-4ccf4aef35dc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.028028] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 805.028028] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 805.028693] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c453102-9b7b-477b-89f7-29f6af97381a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.036401] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2e4f8f69-291f-415d-8c69-713c7879d0b7 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Lock "734e8a2b-faaa-4165-b0e9-69eeb3699b2d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.911s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.036401] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 805.036401] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521ce653-f461-f7da-bf31-345d4d7a88c4" [ 805.036401] env[62600]: _type = "Task" [ 805.036401] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.051962] env[62600]: DEBUG oslo_vmware.api [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222361, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.057639] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521ce653-f461-f7da-bf31-345d4d7a88c4, 'name': SearchDatastore_Task, 'duration_secs': 0.014157} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.058654] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32f55843-4916-4bf7-83b3-c0c3e25b7d7a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.064220] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 805.064220] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ddcf27-b08a-7440-460a-065e40b12119" [ 805.064220] env[62600]: _type = "Task" [ 805.064220] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.072160] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ddcf27-b08a-7440-460a-065e40b12119, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.118882] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc4b631-3d22-485c-8bc3-2b873efeada2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.126025] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6029d55c-54d5-4fc0-bdf2-165be829f86c tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Doing hard reboot of VM {{(pid=62600) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 805.126234] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-a07c864a-fd1f-4445-897f-1c4fd2ef8945 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.131875] env[62600]: DEBUG oslo_vmware.api [None req-6029d55c-54d5-4fc0-bdf2-165be829f86c tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Waiting for the task: (returnval){ [ 805.131875] env[62600]: value = "task-1222363" [ 805.131875] env[62600]: _type = "Task" [ 805.131875] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.139492] env[62600]: DEBUG oslo_vmware.api [None req-6029d55c-54d5-4fc0-bdf2-165be829f86c tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': task-1222363, 'name': ResetVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.147164] env[62600]: DEBUG nova.network.neutron [req-22dcd0ca-97b2-41e8-ad81-88cbd993e9cd req-2c016f20-1786-4eac-b8d6-d617ed5692cb service nova] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Updated VIF entry in instance network info cache for port b4b6687a-b518-40c1-99a5-7d1ffe49e1f2. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 805.147475] env[62600]: DEBUG nova.network.neutron [req-22dcd0ca-97b2-41e8-ad81-88cbd993e9cd req-2c016f20-1786-4eac-b8d6-d617ed5692cb service nova] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Updating instance_info_cache with network_info: [{"id": "b4b6687a-b518-40c1-99a5-7d1ffe49e1f2", "address": "fa:16:3e:2a:79:1c", "network": {"id": "102eb76e-a1ff-4b27-b285-516053dcad35", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-543226554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "888577fb598f45c881757ce3ee52bfa3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4b6687a-b5", "ovs_interfaceid": "b4b6687a-b518-40c1-99a5-7d1ffe49e1f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.155969] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Creating Snapshot of the VM instance {{(pid=62600) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 805.157057] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9105db79-6107-4705-9606-f70b9f854862 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.167118] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': task-1222362, 'name': Rename_Task, 'duration_secs': 0.1409} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.168341] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 805.168653] env[62600]: DEBUG oslo_vmware.api [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 805.168653] env[62600]: value = "task-1222364" [ 805.168653] env[62600]: _type = "Task" [ 805.168653] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.168859] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3cdff250-e1e3-4939-ba07-c1b1ffb59ecb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.177278] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Waiting for the task: (returnval){ [ 805.177278] env[62600]: value = "task-1222365" [ 805.177278] env[62600]: _type = "Task" [ 805.177278] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.180654] env[62600]: DEBUG oslo_vmware.api [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222364, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.188618] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': task-1222365, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.329488] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance af7036df-b9f0-4ce6-962a-1edd7c1ea211 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 805.360523] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Releasing lock "refresh_cache-d34c098d-a827-4dbd-96e4-e27d9d56b847" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.361160] env[62600]: DEBUG nova.compute.manager [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 805.361397] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 805.362375] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-121a9ca8-0a7f-492b-9eeb-97883a8b634e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.370937] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 805.371253] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-85d688ef-6f8d-447c-b2fd-868cd7917d5e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.377620] env[62600]: DEBUG oslo_vmware.api [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 805.377620] env[62600]: value = "task-1222366" [ 805.377620] env[62600]: _type = "Task" [ 805.377620] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.386741] env[62600]: DEBUG oslo_vmware.api [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222366, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.427744] env[62600]: DEBUG nova.network.neutron [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 805.489257] env[62600]: DEBUG nova.objects.instance [None req-aa44dedc-e4b2-43ad-8fca-61e0bc902df2 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Lazy-loading 'pci_requests' on Instance uuid 734e8a2b-faaa-4165-b0e9-69eeb3699b2d {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 805.541359] env[62600]: DEBUG nova.compute.manager [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 805.555662] env[62600]: DEBUG oslo_vmware.api [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222361, 'name': PowerOnVM_Task, 'duration_secs': 0.525281} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.555940] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 805.556148] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-86708a9c-7d63-4ec9-a2ff-13acffb3b27b tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Updating instance 'de4f9304-3357-4eaa-9c94-fe28bc554086' progress to 100 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 805.574396] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ddcf27-b08a-7440-460a-065e40b12119, 'name': SearchDatastore_Task, 'duration_secs': 0.009356} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.574591] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.574843] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] afe838cc-e086-4986-87ec-4e1266bcaf60/afe838cc-e086-4986-87ec-4e1266bcaf60.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 805.575681] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2eb8bf84-ae2f-4b85-ac64-e2b0ffcea04b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.583382] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 805.583382] env[62600]: value = "task-1222367" [ 805.583382] env[62600]: _type = "Task" [ 805.583382] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.587763] env[62600]: DEBUG nova.network.neutron [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Updating instance_info_cache with network_info: [{"id": "1542d23a-48d6-479e-af46-e528ef96651f", "address": "fa:16:3e:48:cf:f1", "network": {"id": "a173b1b7-d062-4f1e-bb7e-be99997291be", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1378511323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a393dbb43124b0f824ffea7665ab55b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1542d23a-48", "ovs_interfaceid": "1542d23a-48d6-479e-af46-e528ef96651f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.591946] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222367, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.642637] env[62600]: DEBUG oslo_vmware.api [None req-6029d55c-54d5-4fc0-bdf2-165be829f86c tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': task-1222363, 'name': ResetVM_Task, 'duration_secs': 0.094093} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.642785] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6029d55c-54d5-4fc0-bdf2-165be829f86c tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Did hard reboot of VM {{(pid=62600) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 805.642980] env[62600]: DEBUG nova.compute.manager [None req-6029d55c-54d5-4fc0-bdf2-165be829f86c tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 805.643748] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b61188-a228-42a0-892b-7b43a31149c5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.650096] env[62600]: DEBUG oslo_concurrency.lockutils [req-22dcd0ca-97b2-41e8-ad81-88cbd993e9cd req-2c016f20-1786-4eac-b8d6-d617ed5692cb service nova] Releasing lock "refresh_cache-afe838cc-e086-4986-87ec-4e1266bcaf60" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.680897] env[62600]: DEBUG oslo_vmware.api [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222364, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.689231] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': task-1222365, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.833210] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance c2c20e5f-6c0e-4989-bc36-ff4a27d5c534 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 805.888412] env[62600]: DEBUG oslo_vmware.api [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222366, 'name': PowerOffVM_Task, 'duration_secs': 0.180361} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.888701] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 805.888911] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 805.889588] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-753b39ab-60b3-49c9-b120-6b486c741765 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.918370] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 805.918569] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 805.918746] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Deleting the datastore file [datastore1] d34c098d-a827-4dbd-96e4-e27d9d56b847 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 805.919037] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1531198-04e8-455c-9e84-18b8add3956b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.926185] env[62600]: DEBUG oslo_vmware.api [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 805.926185] env[62600]: value = "task-1222369" [ 805.926185] env[62600]: _type = "Task" [ 805.926185] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.936825] env[62600]: DEBUG oslo_vmware.api [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222369, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.992067] env[62600]: DEBUG nova.objects.base [None req-aa44dedc-e4b2-43ad-8fca-61e0bc902df2 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Object Instance<734e8a2b-faaa-4165-b0e9-69eeb3699b2d> lazy-loaded attributes: flavor,pci_requests {{(pid=62600) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 805.992390] env[62600]: DEBUG nova.network.neutron [None req-aa44dedc-e4b2-43ad-8fca-61e0bc902df2 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 806.095548] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Releasing lock "refresh_cache-8f3e87a5-bb70-4e50-9fea-6bf53946b8dd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.095994] env[62600]: DEBUG nova.compute.manager [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Instance network_info: |[{"id": "1542d23a-48d6-479e-af46-e528ef96651f", "address": "fa:16:3e:48:cf:f1", "network": {"id": "a173b1b7-d062-4f1e-bb7e-be99997291be", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1378511323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a393dbb43124b0f824ffea7665ab55b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1542d23a-48", "ovs_interfaceid": "1542d23a-48d6-479e-af46-e528ef96651f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 806.096407] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222367, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.097255] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:48:cf:f1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd829efb7-e98e-4b67-bd03-b0888287dbfd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1542d23a-48d6-479e-af46-e528ef96651f', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 806.105066] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Creating folder: Project (7a393dbb43124b0f824ffea7665ab55b). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 806.105800] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-225e3147-8849-4788-af5e-5ae2576bca55 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.108855] env[62600]: DEBUG oslo_concurrency.lockutils [None req-aa44dedc-e4b2-43ad-8fca-61e0bc902df2 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Lock "interface-734e8a2b-faaa-4165-b0e9-69eeb3699b2d-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.130s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.119059] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Created folder: Project (7a393dbb43124b0f824ffea7665ab55b) in parent group-v264198. [ 806.119304] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Creating folder: Instances. Parent ref: group-v264253. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 806.119524] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e5504cee-6741-4913-b6ee-e33961c048b3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.129383] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Created folder: Instances in parent group-v264253. [ 806.129636] env[62600]: DEBUG oslo.service.loopingcall [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.129829] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 806.130054] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f57ce569-df81-47a7-aa99-fc42dc44ec28 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.152223] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 806.152223] env[62600]: value = "task-1222372" [ 806.152223] env[62600]: _type = "Task" [ 806.152223] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.159078] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.160212] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6029d55c-54d5-4fc0-bdf2-165be829f86c tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Lock "92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.933s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.164384] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222372, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.181106] env[62600]: DEBUG oslo_vmware.api [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222364, 'name': CreateSnapshot_Task, 'duration_secs': 0.82705} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.184127] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Created Snapshot of the VM instance {{(pid=62600) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 806.184843] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4b188f6-ff14-4bae-8529-f3e918e1356d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.192269] env[62600]: DEBUG oslo_vmware.api [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': task-1222365, 'name': PowerOnVM_Task, 'duration_secs': 0.643211} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.195317] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 806.195539] env[62600]: INFO nova.compute.manager [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Took 8.24 seconds to spawn the instance on the hypervisor. [ 806.195726] env[62600]: DEBUG nova.compute.manager [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 806.198799] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d6aa134-6cb9-4a23-b026-e48a6aeac9d1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.336807] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 9d899d96-9f4f-41d1-a368-3fde5efc110a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 806.396533] env[62600]: DEBUG nova.compute.manager [req-8fb923f8-35b2-4923-92a8-d5e18d86d632 req-c302ff55-3164-4171-9cd0-98d46b2fe9be service nova] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Received event network-changed-1542d23a-48d6-479e-af46-e528ef96651f {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.396845] env[62600]: DEBUG nova.compute.manager [req-8fb923f8-35b2-4923-92a8-d5e18d86d632 req-c302ff55-3164-4171-9cd0-98d46b2fe9be service nova] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Refreshing instance network info cache due to event network-changed-1542d23a-48d6-479e-af46-e528ef96651f. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 806.397118] env[62600]: DEBUG oslo_concurrency.lockutils [req-8fb923f8-35b2-4923-92a8-d5e18d86d632 req-c302ff55-3164-4171-9cd0-98d46b2fe9be service nova] Acquiring lock "refresh_cache-8f3e87a5-bb70-4e50-9fea-6bf53946b8dd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.397323] env[62600]: DEBUG oslo_concurrency.lockutils [req-8fb923f8-35b2-4923-92a8-d5e18d86d632 req-c302ff55-3164-4171-9cd0-98d46b2fe9be service nova] Acquired lock "refresh_cache-8f3e87a5-bb70-4e50-9fea-6bf53946b8dd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.397537] env[62600]: DEBUG nova.network.neutron [req-8fb923f8-35b2-4923-92a8-d5e18d86d632 req-c302ff55-3164-4171-9cd0-98d46b2fe9be service nova] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Refreshing network info cache for port 1542d23a-48d6-479e-af46-e528ef96651f {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 806.436522] env[62600]: DEBUG oslo_vmware.api [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222369, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.234835} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.436824] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 806.437019] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 806.437215] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 806.437441] env[62600]: INFO nova.compute.manager [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Took 1.08 seconds to destroy the instance on the hypervisor. [ 806.437715] env[62600]: DEBUG oslo.service.loopingcall [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.437922] env[62600]: DEBUG nova.compute.manager [-] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 806.438016] env[62600]: DEBUG nova.network.neutron [-] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 806.454837] env[62600]: DEBUG nova.network.neutron [-] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 806.594849] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222367, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.561572} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.595125] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] afe838cc-e086-4986-87ec-4e1266bcaf60/afe838cc-e086-4986-87ec-4e1266bcaf60.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 806.595357] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 806.595614] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-336dea72-574d-4408-b250-47ec223b3afd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.603362] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 806.603362] env[62600]: value = "task-1222373" [ 806.603362] env[62600]: _type = "Task" [ 806.603362] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.614606] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222373, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.664371] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222372, 'name': CreateVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.709540] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Creating linked-clone VM from snapshot {{(pid=62600) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 806.713647] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-de62a42c-32ab-423d-a7c8-796a231b3995 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.720705] env[62600]: INFO nova.compute.manager [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Took 27.30 seconds to build instance. [ 806.725683] env[62600]: DEBUG oslo_vmware.api [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 806.725683] env[62600]: value = "task-1222374" [ 806.725683] env[62600]: _type = "Task" [ 806.725683] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.733864] env[62600]: DEBUG oslo_vmware.api [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222374, 'name': CloneVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.842220] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance ee7175b9-dbe6-4f90-bd2b-8829194dc6c3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 806.957408] env[62600]: DEBUG nova.network.neutron [-] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.114561] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222373, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.147209} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.114873] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 807.115700] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04712925-3619-4c4e-bcb3-3623bcbed359 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.142537] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] afe838cc-e086-4986-87ec-4e1266bcaf60/afe838cc-e086-4986-87ec-4e1266bcaf60.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 807.143752] env[62600]: DEBUG nova.network.neutron [req-8fb923f8-35b2-4923-92a8-d5e18d86d632 req-c302ff55-3164-4171-9cd0-98d46b2fe9be service nova] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Updated VIF entry in instance network info cache for port 1542d23a-48d6-479e-af46-e528ef96651f. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 807.144082] env[62600]: DEBUG nova.network.neutron [req-8fb923f8-35b2-4923-92a8-d5e18d86d632 req-c302ff55-3164-4171-9cd0-98d46b2fe9be service nova] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Updating instance_info_cache with network_info: [{"id": "1542d23a-48d6-479e-af46-e528ef96651f", "address": "fa:16:3e:48:cf:f1", "network": {"id": "a173b1b7-d062-4f1e-bb7e-be99997291be", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1378511323-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a393dbb43124b0f824ffea7665ab55b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d829efb7-e98e-4b67-bd03-b0888287dbfd", "external-id": "nsx-vlan-transportzone-128", "segmentation_id": 128, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1542d23a-48", "ovs_interfaceid": "1542d23a-48d6-479e-af46-e528ef96651f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.146033] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd2a4e51-2a20-4941-8ce3-e58815401e2f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.171064] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222372, 'name': CreateVM_Task, 'duration_secs': 0.57351} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.172326] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 807.172701] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 807.172701] env[62600]: value = "task-1222375" [ 807.172701] env[62600]: _type = "Task" [ 807.172701] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.173424] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.173607] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.173968] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 807.174314] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40360ee0-d0b0-4019-9b97-19255f25e853 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.183057] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Waiting for the task: (returnval){ [ 807.183057] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52827fe6-ab4e-4118-ac42-e31fa4cde926" [ 807.183057] env[62600]: _type = "Task" [ 807.183057] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.186586] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222375, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.195440] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52827fe6-ab4e-4118-ac42-e31fa4cde926, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.222949] env[62600]: DEBUG oslo_concurrency.lockutils [None req-689c2590-1324-4ea6-b791-39b0ac7e976f tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Lock "58bf5b37-05dd-478f-8c0b-5475ab63bf71" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.887s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.239485] env[62600]: DEBUG oslo_vmware.api [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222374, 'name': CloneVM_Task} progress is 94%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.345468] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance b967fb11-e70e-4e17-b769-38da581bd83b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 807.459837] env[62600]: INFO nova.compute.manager [-] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Took 1.02 seconds to deallocate network for instance. [ 807.492442] env[62600]: DEBUG oslo_concurrency.lockutils [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Acquiring lock "58bf5b37-05dd-478f-8c0b-5475ab63bf71" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.492733] env[62600]: DEBUG oslo_concurrency.lockutils [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Lock "58bf5b37-05dd-478f-8c0b-5475ab63bf71" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.492971] env[62600]: DEBUG oslo_concurrency.lockutils [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Acquiring lock "58bf5b37-05dd-478f-8c0b-5475ab63bf71-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.493213] env[62600]: DEBUG oslo_concurrency.lockutils [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Lock "58bf5b37-05dd-478f-8c0b-5475ab63bf71-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.493670] env[62600]: DEBUG oslo_concurrency.lockutils [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Lock "58bf5b37-05dd-478f-8c0b-5475ab63bf71-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.495923] env[62600]: INFO nova.compute.manager [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Terminating instance [ 807.497955] env[62600]: DEBUG nova.compute.manager [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 807.498198] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 807.499155] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-589c1d62-12d8-406b-af30-660c7b1d4270 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.509818] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 807.510372] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41542f4c-0357-4618-bcd0-c17c51c81424 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.517077] env[62600]: DEBUG oslo_vmware.api [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Waiting for the task: (returnval){ [ 807.517077] env[62600]: value = "task-1222376" [ 807.517077] env[62600]: _type = "Task" [ 807.517077] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.526447] env[62600]: DEBUG oslo_vmware.api [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': task-1222376, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.661860] env[62600]: DEBUG oslo_concurrency.lockutils [req-8fb923f8-35b2-4923-92a8-d5e18d86d632 req-c302ff55-3164-4171-9cd0-98d46b2fe9be service nova] Releasing lock "refresh_cache-8f3e87a5-bb70-4e50-9fea-6bf53946b8dd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.684893] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222375, 'name': ReconfigVM_Task, 'duration_secs': 0.509629} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.685355] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Reconfigured VM instance instance-00000037 to attach disk [datastore2] afe838cc-e086-4986-87ec-4e1266bcaf60/afe838cc-e086-4986-87ec-4e1266bcaf60.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 807.686115] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-63174cc5-681c-4df3-8cf7-a41ed9a1de2d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.698139] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52827fe6-ab4e-4118-ac42-e31fa4cde926, 'name': SearchDatastore_Task, 'duration_secs': 0.01193} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.699729] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.700077] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 807.700297] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.700470] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.700690] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 807.701020] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 807.701020] env[62600]: value = "task-1222377" [ 807.701020] env[62600]: _type = "Task" [ 807.701020] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.701216] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-949c95f4-cbaa-4c92-9dd8-551156de0244 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.711600] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222377, 'name': Rename_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.725762] env[62600]: DEBUG nova.compute.manager [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 807.729218] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 807.729413] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 807.730795] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b9ea99a-8e73-489f-8b4b-f9c025e53730 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.739958] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Waiting for the task: (returnval){ [ 807.739958] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52070e15-f965-7f51-e692-c5ec0a0137c0" [ 807.739958] env[62600]: _type = "Task" [ 807.739958] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.743159] env[62600]: DEBUG oslo_vmware.api [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222374, 'name': CloneVM_Task} progress is 94%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.751695] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52070e15-f965-7f51-e692-c5ec0a0137c0, 'name': SearchDatastore_Task, 'duration_secs': 0.009882} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.752587] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c20c149-05f4-4e1f-88f7-3244658c6bb5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.758821] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Waiting for the task: (returnval){ [ 807.758821] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5285acc8-20af-0008-b53b-22c1f5994d0b" [ 807.758821] env[62600]: _type = "Task" [ 807.758821] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.769796] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5285acc8-20af-0008-b53b-22c1f5994d0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.848511] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance d66c404d-8fd1-4fb7-a3b9-f21854c7e735 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 807.848722] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Migration a871b6f5-809b-4c75-97f7-3f72b2a3dc6a is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 807.848944] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance de4f9304-3357-4eaa-9c94-fe28bc554086 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 807.901072] env[62600]: DEBUG nova.compute.manager [req-3e5012ec-8343-41e0-9fb2-15ac6e19d2a2 req-9e9a446c-b5ca-4a25-bcf3-2c616104e07a service nova] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Received event network-changed-cb1945e8-09df-4f1f-ba94-77e84b24f494 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 807.901401] env[62600]: DEBUG nova.compute.manager [req-3e5012ec-8343-41e0-9fb2-15ac6e19d2a2 req-9e9a446c-b5ca-4a25-bcf3-2c616104e07a service nova] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Refreshing instance network info cache due to event network-changed-cb1945e8-09df-4f1f-ba94-77e84b24f494. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 807.901533] env[62600]: DEBUG oslo_concurrency.lockutils [req-3e5012ec-8343-41e0-9fb2-15ac6e19d2a2 req-9e9a446c-b5ca-4a25-bcf3-2c616104e07a service nova] Acquiring lock "refresh_cache-92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.901723] env[62600]: DEBUG oslo_concurrency.lockutils [req-3e5012ec-8343-41e0-9fb2-15ac6e19d2a2 req-9e9a446c-b5ca-4a25-bcf3-2c616104e07a service nova] Acquired lock "refresh_cache-92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.901907] env[62600]: DEBUG nova.network.neutron [req-3e5012ec-8343-41e0-9fb2-15ac6e19d2a2 req-9e9a446c-b5ca-4a25-bcf3-2c616104e07a service nova] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Refreshing network info cache for port cb1945e8-09df-4f1f-ba94-77e84b24f494 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 807.969029] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.028190] env[62600]: DEBUG oslo_vmware.api [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': task-1222376, 'name': PowerOffVM_Task, 'duration_secs': 0.197356} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.028495] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 808.028671] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 808.028962] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4c71a682-261c-459c-b8bf-3cb68222b412 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.098989] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 808.099166] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 808.099531] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Deleting the datastore file [datastore1] 58bf5b37-05dd-478f-8c0b-5475ab63bf71 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 808.099822] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4bd622c-30b8-41ed-9519-a17d024babf8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.107153] env[62600]: DEBUG oslo_vmware.api [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Waiting for the task: (returnval){ [ 808.107153] env[62600]: value = "task-1222379" [ 808.107153] env[62600]: _type = "Task" [ 808.107153] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.116700] env[62600]: DEBUG oslo_vmware.api [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': task-1222379, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.213443] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222377, 'name': Rename_Task, 'duration_secs': 0.222835} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.213735] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 808.213992] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5bc441a-c13d-4b22-a134-3e4a9cafc1e6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.220227] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 808.220227] env[62600]: value = "task-1222380" [ 808.220227] env[62600]: _type = "Task" [ 808.220227] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.227546] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222380, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.245261] env[62600]: DEBUG oslo_vmware.api [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222374, 'name': CloneVM_Task, 'duration_secs': 1.417884} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.245569] env[62600]: INFO nova.virt.vmwareapi.vmops [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Created linked-clone VM from snapshot [ 808.246436] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aad2dae-6c77-4ea3-a412-ba63dfbcac5d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.252067] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.254953] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Uploading image c1df5cc3-007e-451c-acb1-c9742117a87c {{(pid=62600) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 808.269075] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5285acc8-20af-0008-b53b-22c1f5994d0b, 'name': SearchDatastore_Task, 'duration_secs': 0.010059} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.269362] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.269627] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd/8f3e87a5-bb70-4e50-9fea-6bf53946b8dd.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 808.269889] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-17abfa23-2f8d-401f-9971-56d637a35d27 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.278378] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Waiting for the task: (returnval){ [ 808.278378] env[62600]: value = "task-1222381" [ 808.278378] env[62600]: _type = "Task" [ 808.278378] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.280257] env[62600]: DEBUG oslo_vmware.rw_handles [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 808.280257] env[62600]: value = "vm-264256" [ 808.280257] env[62600]: _type = "VirtualMachine" [ 808.280257] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 808.282330] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1f8a48d4-8ede-4243-8369-e57478fe21ad {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.295841] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': task-1222381, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.297470] env[62600]: DEBUG oslo_vmware.rw_handles [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lease: (returnval){ [ 808.297470] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5272ac24-d58f-8837-8496-664bef345c9f" [ 808.297470] env[62600]: _type = "HttpNfcLease" [ 808.297470] env[62600]: } obtained for exporting VM: (result){ [ 808.297470] env[62600]: value = "vm-264256" [ 808.297470] env[62600]: _type = "VirtualMachine" [ 808.297470] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 808.297807] env[62600]: DEBUG oslo_vmware.api [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the lease: (returnval){ [ 808.297807] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5272ac24-d58f-8837-8496-664bef345c9f" [ 808.297807] env[62600]: _type = "HttpNfcLease" [ 808.297807] env[62600]: } to be ready. {{(pid=62600) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 808.303530] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 808.303530] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5272ac24-d58f-8837-8496-664bef345c9f" [ 808.303530] env[62600]: _type = "HttpNfcLease" [ 808.303530] env[62600]: } is initializing. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 808.351738] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance ddc80377-a14e-48d0-ad24-cc6daf7f6850 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 808.352112] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 808.352197] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3264MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=14 pci_stats=[] {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 808.372035] env[62600]: DEBUG nova.scheduler.client.report [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Refreshing inventories for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 808.388817] env[62600]: DEBUG nova.scheduler.client.report [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Updating ProviderTree inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 808.389127] env[62600]: DEBUG nova.compute.provider_tree [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 808.401998] env[62600]: DEBUG nova.scheduler.client.report [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Refreshing aggregate associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, aggregates: None {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 808.423639] env[62600]: DEBUG nova.network.neutron [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Port fe02d314-372f-43fb-8d7c-cfc33c9e74bd binding to destination host cpu-1 is already ACTIVE {{(pid=62600) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 808.424179] env[62600]: DEBUG oslo_concurrency.lockutils [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "refresh_cache-de4f9304-3357-4eaa-9c94-fe28bc554086" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.424179] env[62600]: DEBUG oslo_concurrency.lockutils [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquired lock "refresh_cache-de4f9304-3357-4eaa-9c94-fe28bc554086" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.424313] env[62600]: DEBUG nova.network.neutron [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 808.426743] env[62600]: DEBUG nova.scheduler.client.report [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Refreshing trait associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 808.472824] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Acquiring lock "734e8a2b-faaa-4165-b0e9-69eeb3699b2d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.473102] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Lock "734e8a2b-faaa-4165-b0e9-69eeb3699b2d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.473354] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Acquiring lock "734e8a2b-faaa-4165-b0e9-69eeb3699b2d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.473626] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Lock "734e8a2b-faaa-4165-b0e9-69eeb3699b2d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.473819] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Lock "734e8a2b-faaa-4165-b0e9-69eeb3699b2d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.475897] env[62600]: INFO nova.compute.manager [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Terminating instance [ 808.481988] env[62600]: DEBUG nova.compute.manager [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 808.482215] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 808.483629] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-332f5af0-8ca4-4328-8545-e0a5d298c45c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.491063] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 808.491309] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c3a44b2-26c2-4059-bcae-a2471d14df68 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.496919] env[62600]: DEBUG oslo_vmware.api [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Waiting for the task: (returnval){ [ 808.496919] env[62600]: value = "task-1222383" [ 808.496919] env[62600]: _type = "Task" [ 808.496919] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.509565] env[62600]: DEBUG oslo_vmware.api [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': task-1222383, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.624330] env[62600]: DEBUG oslo_vmware.api [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Task: {'id': task-1222379, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.333028} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.624330] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 808.624735] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 808.625384] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 808.625800] env[62600]: INFO nova.compute.manager [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Took 1.13 seconds to destroy the instance on the hypervisor. [ 808.626231] env[62600]: DEBUG oslo.service.loopingcall [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.626569] env[62600]: DEBUG nova.compute.manager [-] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 808.629163] env[62600]: DEBUG nova.network.neutron [-] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 808.632585] env[62600]: DEBUG nova.network.neutron [req-3e5012ec-8343-41e0-9fb2-15ac6e19d2a2 req-9e9a446c-b5ca-4a25-bcf3-2c616104e07a service nova] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Updated VIF entry in instance network info cache for port cb1945e8-09df-4f1f-ba94-77e84b24f494. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 808.633240] env[62600]: DEBUG nova.network.neutron [req-3e5012ec-8343-41e0-9fb2-15ac6e19d2a2 req-9e9a446c-b5ca-4a25-bcf3-2c616104e07a service nova] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Updating instance_info_cache with network_info: [{"id": "cb1945e8-09df-4f1f-ba94-77e84b24f494", "address": "fa:16:3e:63:98:19", "network": {"id": "f9decfe0-071c-4ff4-b65a-ea34be9fc487", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-439991991-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "583e67fe6a444511946ce9fac5c66420", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1ce8361b-fd8e-4971-a37f-b84a4f77db19", "external-id": "nsx-vlan-transportzone-255", "segmentation_id": 255, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb1945e8-09", "ovs_interfaceid": "cb1945e8-09df-4f1f-ba94-77e84b24f494", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.731296] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222380, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.791473] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': task-1222381, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.805696] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 808.805696] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5272ac24-d58f-8837-8496-664bef345c9f" [ 808.805696] env[62600]: _type = "HttpNfcLease" [ 808.805696] env[62600]: } is ready. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 808.806297] env[62600]: DEBUG oslo_vmware.rw_handles [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 808.806297] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5272ac24-d58f-8837-8496-664bef345c9f" [ 808.806297] env[62600]: _type = "HttpNfcLease" [ 808.806297] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 808.806801] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf07230d-cce6-45c4-98e4-eaf1ed43ca2a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.818326] env[62600]: DEBUG oslo_vmware.rw_handles [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f55770-edbb-7ba6-4a4c-3bd1fbf4627e/disk-0.vmdk from lease info. {{(pid=62600) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 808.818486] env[62600]: DEBUG oslo_vmware.rw_handles [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f55770-edbb-7ba6-4a4c-3bd1fbf4627e/disk-0.vmdk for reading. {{(pid=62600) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 808.880622] env[62600]: DEBUG oslo_concurrency.lockutils [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquiring lock "92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.880959] env[62600]: DEBUG oslo_concurrency.lockutils [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Lock "92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.881261] env[62600]: DEBUG oslo_concurrency.lockutils [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquiring lock "92bd8b10-2034-49d4-bf38-0edc7e1e6f7d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.881551] env[62600]: DEBUG oslo_concurrency.lockutils [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Lock "92bd8b10-2034-49d4-bf38-0edc7e1e6f7d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.881837] env[62600]: DEBUG oslo_concurrency.lockutils [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Lock "92bd8b10-2034-49d4-bf38-0edc7e1e6f7d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.888270] env[62600]: INFO nova.compute.manager [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Terminating instance [ 808.890815] env[62600]: DEBUG nova.compute.manager [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 808.891010] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 808.891882] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d35b174e-3740-40a8-8cc5-1b7b17795a9f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.900270] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 808.900438] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b27e7d78-4f07-44ab-b1d6-79c1719b3b76 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.910610] env[62600]: DEBUG oslo_vmware.api [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Waiting for the task: (returnval){ [ 808.910610] env[62600]: value = "task-1222384" [ 808.910610] env[62600]: _type = "Task" [ 808.910610] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.918115] env[62600]: DEBUG oslo_vmware.api [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': task-1222384, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.937444] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-66dc55c9-9bce-48bc-86a8-40fcb674dd8e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.015020] env[62600]: DEBUG oslo_vmware.api [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': task-1222383, 'name': PowerOffVM_Task, 'duration_secs': 0.224415} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.015594] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 809.015820] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 809.016674] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1e21bf-6caf-4683-9d04-7b4105b77ed2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.019504] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5d8c405a-7785-4404-927b-6e8ae7ef1c48 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.028107] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c07aace-54b1-44b8-aa37-58406cda5dd6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.067534] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cc04edf-38b6-40c5-ad32-b0ff04ea8c41 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.079017] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c28d4ad0-2ebb-4f81-b287-1d6c21050281 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.092958] env[62600]: DEBUG nova.compute.provider_tree [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 809.097517] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 809.097658] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 809.097887] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Deleting the datastore file [datastore1] 734e8a2b-faaa-4165-b0e9-69eeb3699b2d {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 809.098353] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ae47ed2-058c-416a-a4e9-0fe49e98e580 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.104978] env[62600]: DEBUG oslo_vmware.api [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Waiting for the task: (returnval){ [ 809.104978] env[62600]: value = "task-1222386" [ 809.104978] env[62600]: _type = "Task" [ 809.104978] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.114035] env[62600]: DEBUG oslo_vmware.api [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': task-1222386, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.136540] env[62600]: DEBUG oslo_concurrency.lockutils [req-3e5012ec-8343-41e0-9fb2-15ac6e19d2a2 req-9e9a446c-b5ca-4a25-bcf3-2c616104e07a service nova] Releasing lock "refresh_cache-92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.213569] env[62600]: DEBUG nova.network.neutron [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Updating instance_info_cache with network_info: [{"id": "fe02d314-372f-43fb-8d7c-cfc33c9e74bd", "address": "fa:16:3e:86:22:ec", "network": {"id": "d946b94e-9827-4c22-af05-072fcaa3d0e6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.235", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ec39f2547f624ed0b240d084b949694a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe02d314-37", "ovs_interfaceid": "fe02d314-372f-43fb-8d7c-cfc33c9e74bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.232637] env[62600]: DEBUG oslo_vmware.api [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222380, 'name': PowerOnVM_Task, 'duration_secs': 0.842667} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.233268] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 809.233516] env[62600]: INFO nova.compute.manager [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Took 8.08 seconds to spawn the instance on the hypervisor. [ 809.233701] env[62600]: DEBUG nova.compute.manager [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 809.234652] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34df1f6a-9a73-474f-9157-d0d47939dd6f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.265479] env[62600]: DEBUG nova.compute.manager [req-52f7fbe1-f05a-4bb6-8330-67f939896649 req-325a998f-9fef-4109-a17a-176a79478503 service nova] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Received event network-vif-deleted-73800cd5-a9c3-41ae-b4fe-fbb4d32081b7 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 809.265479] env[62600]: INFO nova.compute.manager [req-52f7fbe1-f05a-4bb6-8330-67f939896649 req-325a998f-9fef-4109-a17a-176a79478503 service nova] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Neutron deleted interface 73800cd5-a9c3-41ae-b4fe-fbb4d32081b7; detaching it from the instance and deleting it from the info cache [ 809.265479] env[62600]: DEBUG nova.network.neutron [req-52f7fbe1-f05a-4bb6-8330-67f939896649 req-325a998f-9fef-4109-a17a-176a79478503 service nova] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.292765] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': task-1222381, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.609203} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.293020] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd/8f3e87a5-bb70-4e50-9fea-6bf53946b8dd.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 809.293226] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 809.293495] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-96f98654-8cb4-493a-8bfc-840af7b9b6ff {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.301046] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Waiting for the task: (returnval){ [ 809.301046] env[62600]: value = "task-1222387" [ 809.301046] env[62600]: _type = "Task" [ 809.301046] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.310031] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': task-1222387, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.416660] env[62600]: DEBUG oslo_vmware.api [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': task-1222384, 'name': PowerOffVM_Task, 'duration_secs': 0.251491} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.417064] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 809.417364] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 809.417696] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd0a588a-6c47-4f66-904e-2d5b24d7e998 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.419472] env[62600]: DEBUG nova.network.neutron [-] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.490172] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 809.490492] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 809.490700] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Deleting the datastore file [datastore1] 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 809.490970] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dea8298b-32b0-4320-8aa9-9aeb7baeca02 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.498066] env[62600]: DEBUG oslo_vmware.api [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Waiting for the task: (returnval){ [ 809.498066] env[62600]: value = "task-1222389" [ 809.498066] env[62600]: _type = "Task" [ 809.498066] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.509429] env[62600]: DEBUG oslo_vmware.api [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': task-1222389, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.616447] env[62600]: DEBUG oslo_vmware.api [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Task: {'id': task-1222386, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.272977} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.616447] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 809.616447] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 809.616629] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 809.616878] env[62600]: INFO nova.compute.manager [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 809.617191] env[62600]: DEBUG oslo.service.loopingcall [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.617395] env[62600]: DEBUG nova.compute.manager [-] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 809.618788] env[62600]: DEBUG nova.network.neutron [-] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 809.638280] env[62600]: DEBUG nova.scheduler.client.report [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 76 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 809.638520] env[62600]: DEBUG nova.compute.provider_tree [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 76 to 77 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 809.638672] env[62600]: DEBUG nova.compute.provider_tree [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 809.718481] env[62600]: DEBUG oslo_concurrency.lockutils [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Releasing lock "refresh_cache-de4f9304-3357-4eaa-9c94-fe28bc554086" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.753109] env[62600]: INFO nova.compute.manager [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Took 27.44 seconds to build instance. [ 809.769839] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aa906390-0356-4b9e-a445-690c0c7ffb54 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.780253] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77fd6c1-e2c7-4b9d-ba6e-c82252268e09 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.810777] env[62600]: DEBUG nova.compute.manager [req-52f7fbe1-f05a-4bb6-8330-67f939896649 req-325a998f-9fef-4109-a17a-176a79478503 service nova] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Detach interface failed, port_id=73800cd5-a9c3-41ae-b4fe-fbb4d32081b7, reason: Instance 58bf5b37-05dd-478f-8c0b-5475ab63bf71 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 809.819408] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': task-1222387, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079937} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.819768] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 809.820622] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-799beb0f-01e2-4719-9d9c-8cec3b96522e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.843254] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd/8f3e87a5-bb70-4e50-9fea-6bf53946b8dd.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 809.843696] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-969178b3-d36c-4b84-a19e-80f04c2fac61 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.863070] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Waiting for the task: (returnval){ [ 809.863070] env[62600]: value = "task-1222390" [ 809.863070] env[62600]: _type = "Task" [ 809.863070] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.871181] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': task-1222390, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.923062] env[62600]: INFO nova.compute.manager [-] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Took 1.30 seconds to deallocate network for instance. [ 810.007983] env[62600]: DEBUG oslo_vmware.api [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Task: {'id': task-1222389, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.352579} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.009038] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 810.009038] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 810.009038] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 810.009176] env[62600]: INFO nova.compute.manager [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Took 1.12 seconds to destroy the instance on the hypervisor. [ 810.009378] env[62600]: DEBUG oslo.service.loopingcall [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 810.009646] env[62600]: DEBUG nova.compute.manager [-] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 810.009752] env[62600]: DEBUG nova.network.neutron [-] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 810.149014] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62600) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 810.149481] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 8.404s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.149775] env[62600]: DEBUG oslo_concurrency.lockutils [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.309s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.155478] env[62600]: INFO nova.compute.claims [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 810.156429] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 810.156429] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Cleaning up deleted instances {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 810.222297] env[62600]: DEBUG nova.compute.manager [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62600) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 810.222297] env[62600]: DEBUG oslo_concurrency.lockutils [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.255731] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1450a0c4-30e2-4209-845a-00ae0d991feb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "afe838cc-e086-4986-87ec-4e1266bcaf60" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.159s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.373536] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': task-1222390, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.402405] env[62600]: DEBUG nova.network.neutron [-] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.431916] env[62600]: DEBUG oslo_concurrency.lockutils [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.667479] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] There are 2 instances to clean {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 810.667703] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 8c0e3a15-0cdd-422a-8a97-95d57621de88] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 810.748228] env[62600]: DEBUG nova.network.neutron [-] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.758666] env[62600]: DEBUG nova.compute.manager [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 810.874575] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': task-1222390, 'name': ReconfigVM_Task, 'duration_secs': 0.665366} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.875052] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Reconfigured VM instance instance-00000038 to attach disk [datastore2] 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd/8f3e87a5-bb70-4e50-9fea-6bf53946b8dd.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 810.875841] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3fa5d7b3-1163-4e85-b1d7-79fecfdb8458 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.882280] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Waiting for the task: (returnval){ [ 810.882280] env[62600]: value = "task-1222391" [ 810.882280] env[62600]: _type = "Task" [ 810.882280] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.890848] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': task-1222391, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.906544] env[62600]: INFO nova.compute.manager [-] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Took 1.29 seconds to deallocate network for instance. [ 811.063503] env[62600]: DEBUG nova.compute.manager [req-970bb42e-a4d6-41d5-be88-c1ee804b1117 req-bd49ba0b-d84d-4d6e-989d-69fef960d38c service nova] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Received event network-changed-b4b6687a-b518-40c1-99a5-7d1ffe49e1f2 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 811.063809] env[62600]: DEBUG nova.compute.manager [req-970bb42e-a4d6-41d5-be88-c1ee804b1117 req-bd49ba0b-d84d-4d6e-989d-69fef960d38c service nova] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Refreshing instance network info cache due to event network-changed-b4b6687a-b518-40c1-99a5-7d1ffe49e1f2. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 811.064067] env[62600]: DEBUG oslo_concurrency.lockutils [req-970bb42e-a4d6-41d5-be88-c1ee804b1117 req-bd49ba0b-d84d-4d6e-989d-69fef960d38c service nova] Acquiring lock "refresh_cache-afe838cc-e086-4986-87ec-4e1266bcaf60" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.064243] env[62600]: DEBUG oslo_concurrency.lockutils [req-970bb42e-a4d6-41d5-be88-c1ee804b1117 req-bd49ba0b-d84d-4d6e-989d-69fef960d38c service nova] Acquired lock "refresh_cache-afe838cc-e086-4986-87ec-4e1266bcaf60" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.064441] env[62600]: DEBUG nova.network.neutron [req-970bb42e-a4d6-41d5-be88-c1ee804b1117 req-bd49ba0b-d84d-4d6e-989d-69fef960d38c service nova] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Refreshing network info cache for port b4b6687a-b518-40c1-99a5-7d1ffe49e1f2 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 811.171432] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 98cbcc28-54e0-40de-9feb-6da917806423] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 811.254789] env[62600]: INFO nova.compute.manager [-] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Took 1.24 seconds to deallocate network for instance. [ 811.291103] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.343360] env[62600]: DEBUG nova.compute.manager [req-e1a63908-09e6-4d26-87a9-263162aaff22 req-212c562d-44fc-40a6-aaad-d67b77a95e4f service nova] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Received event network-vif-deleted-ab8eb6fb-6c95-4452-9f98-963a04b0d64d {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 811.343658] env[62600]: DEBUG nova.compute.manager [req-e1a63908-09e6-4d26-87a9-263162aaff22 req-212c562d-44fc-40a6-aaad-d67b77a95e4f service nova] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Received event network-vif-deleted-cb1945e8-09df-4f1f-ba94-77e84b24f494 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 811.394633] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': task-1222391, 'name': Rename_Task, 'duration_secs': 0.2672} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.394964] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 811.395232] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2ae1f354-46a9-47ca-999a-438289735152 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.401408] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Waiting for the task: (returnval){ [ 811.401408] env[62600]: value = "task-1222392" [ 811.401408] env[62600]: _type = "Task" [ 811.401408] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.414888] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.415594] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': task-1222392, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.658835] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6f587b3-edc6-49b3-8416-83f3f2658361 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.668301] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3302c08-6049-4081-805d-35be1270bbea {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.702265] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 811.702465] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Cleaning up deleted instances with incomplete migration {{(pid=62600) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 811.704167] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e875d49-88ba-4458-9d0f-6efbfd5c4e67 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.713514] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9610e57-5ee1-439a-9ad1-82f360f0f153 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.729765] env[62600]: DEBUG nova.compute.provider_tree [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 811.770750] env[62600]: DEBUG oslo_concurrency.lockutils [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.858090] env[62600]: DEBUG nova.network.neutron [req-970bb42e-a4d6-41d5-be88-c1ee804b1117 req-bd49ba0b-d84d-4d6e-989d-69fef960d38c service nova] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Updated VIF entry in instance network info cache for port b4b6687a-b518-40c1-99a5-7d1ffe49e1f2. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 811.858770] env[62600]: DEBUG nova.network.neutron [req-970bb42e-a4d6-41d5-be88-c1ee804b1117 req-bd49ba0b-d84d-4d6e-989d-69fef960d38c service nova] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Updating instance_info_cache with network_info: [{"id": "b4b6687a-b518-40c1-99a5-7d1ffe49e1f2", "address": "fa:16:3e:2a:79:1c", "network": {"id": "102eb76e-a1ff-4b27-b285-516053dcad35", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-543226554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "888577fb598f45c881757ce3ee52bfa3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4b6687a-b5", "ovs_interfaceid": "b4b6687a-b518-40c1-99a5-7d1ffe49e1f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.916358] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': task-1222392, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.208404] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 812.268771] env[62600]: DEBUG nova.scheduler.client.report [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 77 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 812.268898] env[62600]: DEBUG nova.compute.provider_tree [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 77 to 78 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 812.269139] env[62600]: DEBUG nova.compute.provider_tree [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 812.362749] env[62600]: DEBUG oslo_concurrency.lockutils [req-970bb42e-a4d6-41d5-be88-c1ee804b1117 req-bd49ba0b-d84d-4d6e-989d-69fef960d38c service nova] Releasing lock "refresh_cache-afe838cc-e086-4986-87ec-4e1266bcaf60" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.418703] env[62600]: DEBUG oslo_vmware.api [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': task-1222392, 'name': PowerOnVM_Task, 'duration_secs': 0.576415} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.418867] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 812.419092] env[62600]: INFO nova.compute.manager [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Took 8.66 seconds to spawn the instance on the hypervisor. [ 812.419327] env[62600]: DEBUG nova.compute.manager [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 812.420494] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d63c52d-7765-4d90-9639-f3e738f1a15a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.775355] env[62600]: DEBUG oslo_concurrency.lockutils [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.625s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.775878] env[62600]: DEBUG nova.compute.manager [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 812.778943] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.729s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.780441] env[62600]: INFO nova.compute.claims [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 812.937377] env[62600]: INFO nova.compute.manager [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Took 30.14 seconds to build instance. [ 813.285706] env[62600]: DEBUG nova.compute.utils [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 813.289100] env[62600]: DEBUG nova.compute.manager [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 813.289333] env[62600]: DEBUG nova.network.neutron [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 813.343437] env[62600]: DEBUG nova.policy [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fbd1b2fb34d841359ada8fc44bec2986', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b028450f2da445fb83e37adfc86bba68', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 813.440245] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6d5d747f-d276-466f-9076-142fede66c59 tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Lock "8f3e87a5-bb70-4e50-9fea-6bf53946b8dd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.005s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.648145] env[62600]: DEBUG nova.network.neutron [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Successfully created port: 75e67d5c-8153-4a74-b875-19d68778a85a {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 813.789937] env[62600]: DEBUG nova.compute.manager [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 813.943567] env[62600]: DEBUG nova.compute.manager [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 814.182768] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12438f4-c5ff-4075-8c5e-9cfa9b7d0691 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.191356] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97e045c9-9255-47d0-bc43-9dc462c671dc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.224212] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e3cee19-c53a-4d67-bf22-7afc107641e1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.232138] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6884a63d-8ab4-4dc6-adff-d2bed321166f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.247642] env[62600]: DEBUG nova.compute.provider_tree [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 814.462884] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.750740] env[62600]: DEBUG nova.scheduler.client.report [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 814.801714] env[62600]: DEBUG nova.compute.manager [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 814.833597] env[62600]: DEBUG nova.virt.hardware [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 814.833851] env[62600]: DEBUG nova.virt.hardware [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 814.834092] env[62600]: DEBUG nova.virt.hardware [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 814.834344] env[62600]: DEBUG nova.virt.hardware [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 814.834512] env[62600]: DEBUG nova.virt.hardware [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 814.834668] env[62600]: DEBUG nova.virt.hardware [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 814.835814] env[62600]: DEBUG nova.virt.hardware [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 814.835814] env[62600]: DEBUG nova.virt.hardware [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 814.835814] env[62600]: DEBUG nova.virt.hardware [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 814.835814] env[62600]: DEBUG nova.virt.hardware [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 814.835814] env[62600]: DEBUG nova.virt.hardware [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 814.836508] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bbbddc3-7464-417d-9b02-f7af24cc5883 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.845153] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e77a3ef-d443-4cee-9a25-3b071d63c46a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.871892] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Acquiring lock "8f3e87a5-bb70-4e50-9fea-6bf53946b8dd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.872211] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Lock "8f3e87a5-bb70-4e50-9fea-6bf53946b8dd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.872428] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Acquiring lock "8f3e87a5-bb70-4e50-9fea-6bf53946b8dd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.872621] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Lock "8f3e87a5-bb70-4e50-9fea-6bf53946b8dd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.872797] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Lock "8f3e87a5-bb70-4e50-9fea-6bf53946b8dd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.877747] env[62600]: INFO nova.compute.manager [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Terminating instance [ 814.879740] env[62600]: DEBUG nova.compute.manager [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 814.880228] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 814.881141] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d61cfe6e-cd0f-482c-a2d4-2d0a903f8941 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.888634] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 814.888877] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d3fedd55-9849-4d8e-9f85-e6a00b23f9da {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.895574] env[62600]: DEBUG oslo_vmware.api [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Waiting for the task: (returnval){ [ 814.895574] env[62600]: value = "task-1222393" [ 814.895574] env[62600]: _type = "Task" [ 814.895574] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.904448] env[62600]: DEBUG oslo_vmware.api [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': task-1222393, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.229106] env[62600]: DEBUG nova.compute.manager [req-f656be3c-308e-44ab-bbf5-5256c63e8bce req-d06088cb-06a3-4a30-b6f9-2af15fa55757 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Received event network-vif-plugged-75e67d5c-8153-4a74-b875-19d68778a85a {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 815.229397] env[62600]: DEBUG oslo_concurrency.lockutils [req-f656be3c-308e-44ab-bbf5-5256c63e8bce req-d06088cb-06a3-4a30-b6f9-2af15fa55757 service nova] Acquiring lock "bd9b7351-01a2-429d-a860-d9ff6855eefc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.229977] env[62600]: DEBUG oslo_concurrency.lockutils [req-f656be3c-308e-44ab-bbf5-5256c63e8bce req-d06088cb-06a3-4a30-b6f9-2af15fa55757 service nova] Lock "bd9b7351-01a2-429d-a860-d9ff6855eefc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.230875] env[62600]: DEBUG oslo_concurrency.lockutils [req-f656be3c-308e-44ab-bbf5-5256c63e8bce req-d06088cb-06a3-4a30-b6f9-2af15fa55757 service nova] Lock "bd9b7351-01a2-429d-a860-d9ff6855eefc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.231220] env[62600]: DEBUG nova.compute.manager [req-f656be3c-308e-44ab-bbf5-5256c63e8bce req-d06088cb-06a3-4a30-b6f9-2af15fa55757 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] No waiting events found dispatching network-vif-plugged-75e67d5c-8153-4a74-b875-19d68778a85a {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 815.231454] env[62600]: WARNING nova.compute.manager [req-f656be3c-308e-44ab-bbf5-5256c63e8bce req-d06088cb-06a3-4a30-b6f9-2af15fa55757 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Received unexpected event network-vif-plugged-75e67d5c-8153-4a74-b875-19d68778a85a for instance with vm_state building and task_state spawning. [ 815.256259] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.477s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.256846] env[62600]: DEBUG nova.compute.manager [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 815.259529] env[62600]: DEBUG oslo_concurrency.lockutils [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.878s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.261187] env[62600]: INFO nova.compute.claims [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 815.406121] env[62600]: DEBUG oslo_vmware.api [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': task-1222393, 'name': PowerOffVM_Task, 'duration_secs': 0.208345} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.406408] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 815.406584] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 815.406832] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6c2ce4e9-1d52-445c-8d6b-3ce0b30f43e2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.465249] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 815.465548] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 815.465784] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Deleting the datastore file [datastore2] 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 815.466107] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-95cdfd78-c188-43c5-bb7c-0c6a2e56e01d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.472816] env[62600]: DEBUG oslo_vmware.api [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Waiting for the task: (returnval){ [ 815.472816] env[62600]: value = "task-1222395" [ 815.472816] env[62600]: _type = "Task" [ 815.472816] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.481281] env[62600]: DEBUG oslo_vmware.api [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': task-1222395, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.702314] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Acquiring lock "92cf1acb-b471-4add-a73f-7e9e94fbaaa4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.702463] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Lock "92cf1acb-b471-4add-a73f-7e9e94fbaaa4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.702682] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Acquiring lock "92cf1acb-b471-4add-a73f-7e9e94fbaaa4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.702873] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Lock "92cf1acb-b471-4add-a73f-7e9e94fbaaa4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.703057] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Lock "92cf1acb-b471-4add-a73f-7e9e94fbaaa4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.705233] env[62600]: INFO nova.compute.manager [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Terminating instance [ 815.709037] env[62600]: DEBUG nova.compute.manager [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 815.709295] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 815.709549] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-032e85f7-42ad-47df-a808-2e73b5f93062 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.717812] env[62600]: DEBUG oslo_vmware.api [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Waiting for the task: (returnval){ [ 815.717812] env[62600]: value = "task-1222396" [ 815.717812] env[62600]: _type = "Task" [ 815.717812] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.726650] env[62600]: DEBUG oslo_vmware.api [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222396, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.748485] env[62600]: DEBUG nova.network.neutron [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Successfully updated port: 75e67d5c-8153-4a74-b875-19d68778a85a {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 815.770296] env[62600]: DEBUG nova.compute.utils [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 815.772497] env[62600]: DEBUG nova.compute.manager [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 815.772497] env[62600]: DEBUG nova.network.neutron [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 815.812699] env[62600]: DEBUG nova.compute.manager [req-93f97c36-2bb8-4f0d-982c-236434f2dd55 req-173c9415-300f-4dd1-8ebd-e504d70b07f9 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Received event network-changed-75e67d5c-8153-4a74-b875-19d68778a85a {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 815.813044] env[62600]: DEBUG nova.compute.manager [req-93f97c36-2bb8-4f0d-982c-236434f2dd55 req-173c9415-300f-4dd1-8ebd-e504d70b07f9 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Refreshing instance network info cache due to event network-changed-75e67d5c-8153-4a74-b875-19d68778a85a. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 815.813297] env[62600]: DEBUG oslo_concurrency.lockutils [req-93f97c36-2bb8-4f0d-982c-236434f2dd55 req-173c9415-300f-4dd1-8ebd-e504d70b07f9 service nova] Acquiring lock "refresh_cache-bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.813357] env[62600]: DEBUG oslo_concurrency.lockutils [req-93f97c36-2bb8-4f0d-982c-236434f2dd55 req-173c9415-300f-4dd1-8ebd-e504d70b07f9 service nova] Acquired lock "refresh_cache-bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.813773] env[62600]: DEBUG nova.network.neutron [req-93f97c36-2bb8-4f0d-982c-236434f2dd55 req-173c9415-300f-4dd1-8ebd-e504d70b07f9 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Refreshing network info cache for port 75e67d5c-8153-4a74-b875-19d68778a85a {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 815.825841] env[62600]: DEBUG nova.policy [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '76cb336f935f49d995ea74c6fc8a2bad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '11c88a6316a6409e9ea3d023240c366d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 815.983100] env[62600]: DEBUG oslo_vmware.api [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Task: {'id': task-1222395, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.401563} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.983671] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 815.983873] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 815.984070] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 815.984258] env[62600]: INFO nova.compute.manager [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Took 1.10 seconds to destroy the instance on the hypervisor. [ 815.984535] env[62600]: DEBUG oslo.service.loopingcall [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.984758] env[62600]: DEBUG nova.compute.manager [-] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 815.984851] env[62600]: DEBUG nova.network.neutron [-] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 816.165728] env[62600]: DEBUG nova.network.neutron [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Successfully created port: a1ade93e-f46e-4261-9a3e-45ffbb2aa092 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 816.233190] env[62600]: DEBUG oslo_vmware.api [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222396, 'name': PowerOffVM_Task, 'duration_secs': 0.199414} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.233542] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 816.233755] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Volume detach. Driver type: vmdk {{(pid=62600) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 816.233955] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264214', 'volume_id': '88d3934f-d0ae-4578-af24-9abb842074b3', 'name': 'volume-88d3934f-d0ae-4578-af24-9abb842074b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '92cf1acb-b471-4add-a73f-7e9e94fbaaa4', 'attached_at': '', 'detached_at': '', 'volume_id': '88d3934f-d0ae-4578-af24-9abb842074b3', 'serial': '88d3934f-d0ae-4578-af24-9abb842074b3'} {{(pid=62600) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 816.234789] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f29b3a6-135c-4c06-b9ad-e91848288d5e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.256152] env[62600]: DEBUG oslo_concurrency.lockutils [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "refresh_cache-bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.257158] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e07b1c06-0b5d-4115-b752-b13b22026247 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.264962] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60401ce9-e51b-49da-b2f1-25e3685ea399 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.288718] env[62600]: DEBUG nova.compute.manager [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 816.292543] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6698c2f5-0652-48b1-958c-7be469aed6c3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.310629] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] The volume has not been displaced from its original location: [datastore2] volume-88d3934f-d0ae-4578-af24-9abb842074b3/volume-88d3934f-d0ae-4578-af24-9abb842074b3.vmdk. No consolidation needed. {{(pid=62600) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 816.316745] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Reconfiguring VM instance instance-0000002c to detach disk 2000 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 816.322166] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5829fd3e-727f-49df-84e7-e621b3691e25 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.345943] env[62600]: DEBUG oslo_vmware.api [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Waiting for the task: (returnval){ [ 816.345943] env[62600]: value = "task-1222397" [ 816.345943] env[62600]: _type = "Task" [ 816.345943] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.358534] env[62600]: DEBUG oslo_vmware.api [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222397, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.376369] env[62600]: DEBUG nova.network.neutron [req-93f97c36-2bb8-4f0d-982c-236434f2dd55 req-173c9415-300f-4dd1-8ebd-e504d70b07f9 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 816.596727] env[62600]: DEBUG nova.network.neutron [req-93f97c36-2bb8-4f0d-982c-236434f2dd55 req-173c9415-300f-4dd1-8ebd-e504d70b07f9 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.679293] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d82940d-329e-4008-8fab-86d312aa1d80 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.687849] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-572ba630-982e-4c26-9f9f-80f8e0986aa2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.721028] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b10d902-cd77-4ffb-b06c-c2c418bf85bc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.729368] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e94129c3-b7ce-4627-9005-8fbd0ed13eae {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.743810] env[62600]: DEBUG nova.compute.provider_tree [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.856846] env[62600]: DEBUG oslo_vmware.api [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222397, 'name': ReconfigVM_Task, 'duration_secs': 0.161462} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.857228] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Reconfigured VM instance instance-0000002c to detach disk 2000 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 816.862415] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5dc2c17a-a728-47c2-9788-5b34d0a5ca77 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.879843] env[62600]: DEBUG oslo_vmware.api [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Waiting for the task: (returnval){ [ 816.879843] env[62600]: value = "task-1222398" [ 816.879843] env[62600]: _type = "Task" [ 816.879843] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.888512] env[62600]: DEBUG oslo_vmware.api [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222398, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.922913] env[62600]: DEBUG nova.network.neutron [-] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.099356] env[62600]: DEBUG oslo_concurrency.lockutils [req-93f97c36-2bb8-4f0d-982c-236434f2dd55 req-173c9415-300f-4dd1-8ebd-e504d70b07f9 service nova] Releasing lock "refresh_cache-bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.099721] env[62600]: DEBUG oslo_concurrency.lockutils [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "refresh_cache-bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.099889] env[62600]: DEBUG nova.network.neutron [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 817.247967] env[62600]: DEBUG nova.scheduler.client.report [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 817.302334] env[62600]: DEBUG nova.compute.manager [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 817.325564] env[62600]: DEBUG nova.virt.hardware [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 817.325813] env[62600]: DEBUG nova.virt.hardware [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 817.325974] env[62600]: DEBUG nova.virt.hardware [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 817.326176] env[62600]: DEBUG nova.virt.hardware [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 817.326444] env[62600]: DEBUG nova.virt.hardware [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 817.326502] env[62600]: DEBUG nova.virt.hardware [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 817.326685] env[62600]: DEBUG nova.virt.hardware [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 817.326850] env[62600]: DEBUG nova.virt.hardware [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 817.327098] env[62600]: DEBUG nova.virt.hardware [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 817.327332] env[62600]: DEBUG nova.virt.hardware [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 817.327522] env[62600]: DEBUG nova.virt.hardware [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 817.328516] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f7b1fe-2a99-4733-b243-13bd9f80872c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.336505] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd0fd6f-0fc3-4abf-945c-60899943f5c7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.391033] env[62600]: DEBUG oslo_vmware.api [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222398, 'name': ReconfigVM_Task, 'duration_secs': 0.308227} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.391033] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264214', 'volume_id': '88d3934f-d0ae-4578-af24-9abb842074b3', 'name': 'volume-88d3934f-d0ae-4578-af24-9abb842074b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '92cf1acb-b471-4add-a73f-7e9e94fbaaa4', 'attached_at': '', 'detached_at': '', 'volume_id': '88d3934f-d0ae-4578-af24-9abb842074b3', 'serial': '88d3934f-d0ae-4578-af24-9abb842074b3'} {{(pid=62600) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 817.391033] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 817.391862] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58c59b60-1ae9-4b2f-b3f0-81451a53f0b7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.399299] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 817.399561] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d663078e-cfbd-4cf2-aa71-1cdfb46fdfa5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.426083] env[62600]: INFO nova.compute.manager [-] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Took 1.44 seconds to deallocate network for instance. [ 817.459516] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 817.459798] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 817.460050] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Deleting the datastore file [datastore2] 92cf1acb-b471-4add-a73f-7e9e94fbaaa4 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 817.460413] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-572b827c-2a3f-4d10-bfe7-87ba3c82bba4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.468565] env[62600]: DEBUG oslo_vmware.api [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Waiting for the task: (returnval){ [ 817.468565] env[62600]: value = "task-1222400" [ 817.468565] env[62600]: _type = "Task" [ 817.468565] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.478107] env[62600]: DEBUG oslo_vmware.api [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222400, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.668366] env[62600]: DEBUG nova.network.neutron [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 817.753090] env[62600]: DEBUG oslo_concurrency.lockutils [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.493s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.753699] env[62600]: DEBUG nova.compute.manager [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 817.756717] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.679s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.758169] env[62600]: INFO nova.compute.claims [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 817.809084] env[62600]: DEBUG nova.network.neutron [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Successfully updated port: a1ade93e-f46e-4261-9a3e-45ffbb2aa092 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 817.896995] env[62600]: DEBUG nova.network.neutron [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Updating instance_info_cache with network_info: [{"id": "75e67d5c-8153-4a74-b875-19d68778a85a", "address": "fa:16:3e:fa:9d:b3", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75e67d5c-81", "ovs_interfaceid": "75e67d5c-8153-4a74-b875-19d68778a85a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.934626] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.962951] env[62600]: DEBUG nova.compute.manager [req-2da85701-4cf3-4774-acdb-00015b620c48 req-5bb4d813-6935-46b0-a976-2e8477a78d9e service nova] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Received event network-vif-deleted-1542d23a-48d6-479e-af46-e528ef96651f {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.963209] env[62600]: DEBUG nova.compute.manager [req-2da85701-4cf3-4774-acdb-00015b620c48 req-5bb4d813-6935-46b0-a976-2e8477a78d9e service nova] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Received event network-vif-plugged-a1ade93e-f46e-4261-9a3e-45ffbb2aa092 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.963409] env[62600]: DEBUG oslo_concurrency.lockutils [req-2da85701-4cf3-4774-acdb-00015b620c48 req-5bb4d813-6935-46b0-a976-2e8477a78d9e service nova] Acquiring lock "4e1a376f-6619-4c35-b75f-b45db75815cc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.963615] env[62600]: DEBUG oslo_concurrency.lockutils [req-2da85701-4cf3-4774-acdb-00015b620c48 req-5bb4d813-6935-46b0-a976-2e8477a78d9e service nova] Lock "4e1a376f-6619-4c35-b75f-b45db75815cc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.963842] env[62600]: DEBUG oslo_concurrency.lockutils [req-2da85701-4cf3-4774-acdb-00015b620c48 req-5bb4d813-6935-46b0-a976-2e8477a78d9e service nova] Lock "4e1a376f-6619-4c35-b75f-b45db75815cc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.963960] env[62600]: DEBUG nova.compute.manager [req-2da85701-4cf3-4774-acdb-00015b620c48 req-5bb4d813-6935-46b0-a976-2e8477a78d9e service nova] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] No waiting events found dispatching network-vif-plugged-a1ade93e-f46e-4261-9a3e-45ffbb2aa092 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 817.964146] env[62600]: WARNING nova.compute.manager [req-2da85701-4cf3-4774-acdb-00015b620c48 req-5bb4d813-6935-46b0-a976-2e8477a78d9e service nova] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Received unexpected event network-vif-plugged-a1ade93e-f46e-4261-9a3e-45ffbb2aa092 for instance with vm_state building and task_state spawning. [ 817.964405] env[62600]: DEBUG nova.compute.manager [req-2da85701-4cf3-4774-acdb-00015b620c48 req-5bb4d813-6935-46b0-a976-2e8477a78d9e service nova] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Received event network-changed-a1ade93e-f46e-4261-9a3e-45ffbb2aa092 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.964484] env[62600]: DEBUG nova.compute.manager [req-2da85701-4cf3-4774-acdb-00015b620c48 req-5bb4d813-6935-46b0-a976-2e8477a78d9e service nova] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Refreshing instance network info cache due to event network-changed-a1ade93e-f46e-4261-9a3e-45ffbb2aa092. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 817.964672] env[62600]: DEBUG oslo_concurrency.lockutils [req-2da85701-4cf3-4774-acdb-00015b620c48 req-5bb4d813-6935-46b0-a976-2e8477a78d9e service nova] Acquiring lock "refresh_cache-4e1a376f-6619-4c35-b75f-b45db75815cc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.964816] env[62600]: DEBUG oslo_concurrency.lockutils [req-2da85701-4cf3-4774-acdb-00015b620c48 req-5bb4d813-6935-46b0-a976-2e8477a78d9e service nova] Acquired lock "refresh_cache-4e1a376f-6619-4c35-b75f-b45db75815cc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.964979] env[62600]: DEBUG nova.network.neutron [req-2da85701-4cf3-4774-acdb-00015b620c48 req-5bb4d813-6935-46b0-a976-2e8477a78d9e service nova] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Refreshing network info cache for port a1ade93e-f46e-4261-9a3e-45ffbb2aa092 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 817.978463] env[62600]: DEBUG oslo_vmware.api [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Task: {'id': task-1222400, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.123858} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.978679] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 817.978871] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 817.979059] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 817.979265] env[62600]: INFO nova.compute.manager [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Took 2.27 seconds to destroy the instance on the hypervisor. [ 817.979542] env[62600]: DEBUG oslo.service.loopingcall [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 817.979703] env[62600]: DEBUG nova.compute.manager [-] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 817.979790] env[62600]: DEBUG nova.network.neutron [-] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 818.269069] env[62600]: DEBUG nova.compute.utils [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 818.269069] env[62600]: DEBUG nova.compute.manager [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 818.269069] env[62600]: DEBUG nova.network.neutron [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 818.309288] env[62600]: DEBUG nova.policy [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cb6b500f66c249629e0bb20838dca631', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '88985cba32da4faaa133d74ffa635292', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 818.313814] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Acquiring lock "refresh_cache-4e1a376f-6619-4c35-b75f-b45db75815cc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.400210] env[62600]: DEBUG oslo_concurrency.lockutils [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "refresh_cache-bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.400690] env[62600]: DEBUG nova.compute.manager [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Instance network_info: |[{"id": "75e67d5c-8153-4a74-b875-19d68778a85a", "address": "fa:16:3e:fa:9d:b3", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75e67d5c-81", "ovs_interfaceid": "75e67d5c-8153-4a74-b875-19d68778a85a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 818.401206] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fa:9d:b3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '184687d6-125a-4b58-bb5b-fdb404088eda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '75e67d5c-8153-4a74-b875-19d68778a85a', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 818.408901] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Creating folder: Project (b028450f2da445fb83e37adfc86bba68). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 818.409273] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5b267703-d45b-4489-a9f4-0c5e42ce0d77 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.420608] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Created folder: Project (b028450f2da445fb83e37adfc86bba68) in parent group-v264198. [ 818.420812] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Creating folder: Instances. Parent ref: group-v264257. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 818.421069] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7d08c304-ca86-4837-a623-066e79e59e64 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.430386] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Created folder: Instances in parent group-v264257. [ 818.430632] env[62600]: DEBUG oslo.service.loopingcall [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 818.430833] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 818.431061] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-75ea7cb5-b9dc-4d7a-9991-f0ebc88b2b83 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.450508] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 818.450508] env[62600]: value = "task-1222403" [ 818.450508] env[62600]: _type = "Task" [ 818.450508] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.458480] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222403, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.548322] env[62600]: DEBUG nova.network.neutron [req-2da85701-4cf3-4774-acdb-00015b620c48 req-5bb4d813-6935-46b0-a976-2e8477a78d9e service nova] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 818.675700] env[62600]: DEBUG nova.network.neutron [req-2da85701-4cf3-4774-acdb-00015b620c48 req-5bb4d813-6935-46b0-a976-2e8477a78d9e service nova] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.776761] env[62600]: DEBUG nova.compute.manager [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 818.866190] env[62600]: DEBUG nova.network.neutron [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Successfully created port: 6832182f-fe16-4367-a4bc-53d7bd8f6033 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 818.968785] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222403, 'name': CreateVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.995556] env[62600]: DEBUG oslo_vmware.rw_handles [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f55770-edbb-7ba6-4a4c-3bd1fbf4627e/disk-0.vmdk. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 818.996498] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a28c6b-21f7-4f71-ad1f-fb979602d9e1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.006277] env[62600]: DEBUG oslo_vmware.rw_handles [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f55770-edbb-7ba6-4a4c-3bd1fbf4627e/disk-0.vmdk is in state: ready. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 819.006463] env[62600]: ERROR oslo_vmware.rw_handles [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f55770-edbb-7ba6-4a4c-3bd1fbf4627e/disk-0.vmdk due to incomplete transfer. [ 819.006703] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-145eecbf-96fe-4131-a39d-195e64e5a88e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.016632] env[62600]: DEBUG oslo_vmware.rw_handles [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52f55770-edbb-7ba6-4a4c-3bd1fbf4627e/disk-0.vmdk. {{(pid=62600) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 819.016846] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Uploaded image c1df5cc3-007e-451c-acb1-c9742117a87c to the Glance image server {{(pid=62600) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 819.020292] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Destroying the VM {{(pid=62600) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 819.023235] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-efafad50-f722-4bcc-bc1d-cc47569ff5d4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.030273] env[62600]: DEBUG oslo_vmware.api [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 819.030273] env[62600]: value = "task-1222404" [ 819.030273] env[62600]: _type = "Task" [ 819.030273] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.041339] env[62600]: DEBUG oslo_vmware.api [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222404, 'name': Destroy_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.050811] env[62600]: DEBUG nova.compute.manager [req-6b4b096c-74db-4602-bba6-201367bf79b9 req-599c0791-68d8-4bc6-b673-77c8566577b1 service nova] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Received event network-vif-deleted-779d4e97-daa9-44ec-9ee0-91658a73bc02 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.053758] env[62600]: INFO nova.compute.manager [req-6b4b096c-74db-4602-bba6-201367bf79b9 req-599c0791-68d8-4bc6-b673-77c8566577b1 service nova] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Neutron deleted interface 779d4e97-daa9-44ec-9ee0-91658a73bc02; detaching it from the instance and deleting it from the info cache [ 819.053758] env[62600]: DEBUG nova.network.neutron [req-6b4b096c-74db-4602-bba6-201367bf79b9 req-599c0791-68d8-4bc6-b673-77c8566577b1 service nova] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.181518] env[62600]: DEBUG oslo_concurrency.lockutils [req-2da85701-4cf3-4774-acdb-00015b620c48 req-5bb4d813-6935-46b0-a976-2e8477a78d9e service nova] Releasing lock "refresh_cache-4e1a376f-6619-4c35-b75f-b45db75815cc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.181743] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Acquired lock "refresh_cache-4e1a376f-6619-4c35-b75f-b45db75815cc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.181903] env[62600]: DEBUG nova.network.neutron [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 819.250406] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57392eef-4c26-42ba-8ccf-55fea23602c1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.259276] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8db041c-6fdf-49ec-b936-dc612ed173a8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.301948] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece08c43-8137-4577-80aa-fba01fc5613b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.310213] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04519414-8d54-407b-af8e-4158292e934f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.325104] env[62600]: DEBUG nova.compute.provider_tree [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.437539] env[62600]: DEBUG nova.network.neutron [-] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.461529] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222403, 'name': CreateVM_Task, 'duration_secs': 0.59579} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.461721] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 819.462441] env[62600]: DEBUG oslo_concurrency.lockutils [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.463635] env[62600]: DEBUG oslo_concurrency.lockutils [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.463635] env[62600]: DEBUG oslo_concurrency.lockutils [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 819.463635] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5f41170-294d-4269-85f5-68a7bf76f26b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.467445] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 819.467445] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f5d346-989d-591a-b684-2ad26c34135e" [ 819.467445] env[62600]: _type = "Task" [ 819.467445] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.478124] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f5d346-989d-591a-b684-2ad26c34135e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.540288] env[62600]: DEBUG oslo_vmware.api [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222404, 'name': Destroy_Task, 'duration_secs': 0.358698} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.540577] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Destroyed the VM [ 819.540815] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Deleting Snapshot of the VM instance {{(pid=62600) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 819.541082] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-fd1ed46f-9be4-47d4-85d6-ccc11d093204 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.547301] env[62600]: DEBUG oslo_vmware.api [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 819.547301] env[62600]: value = "task-1222405" [ 819.547301] env[62600]: _type = "Task" [ 819.547301] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.555703] env[62600]: DEBUG oslo_vmware.api [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222405, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.555978] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-05120b7d-e854-4c6a-9b39-531e08ac0adc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.563931] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77cb531c-54be-4538-b08a-5661de8fb008 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.590852] env[62600]: DEBUG nova.compute.manager [req-6b4b096c-74db-4602-bba6-201367bf79b9 req-599c0791-68d8-4bc6-b673-77c8566577b1 service nova] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Detach interface failed, port_id=779d4e97-daa9-44ec-9ee0-91658a73bc02, reason: Instance 92cf1acb-b471-4add-a73f-7e9e94fbaaa4 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 819.738045] env[62600]: DEBUG nova.network.neutron [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 819.802641] env[62600]: DEBUG nova.compute.manager [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 819.829020] env[62600]: DEBUG nova.virt.hardware [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 819.829020] env[62600]: DEBUG nova.virt.hardware [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 819.829020] env[62600]: DEBUG nova.virt.hardware [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 819.829020] env[62600]: DEBUG nova.virt.hardware [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 819.829294] env[62600]: DEBUG nova.virt.hardware [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 819.829294] env[62600]: DEBUG nova.virt.hardware [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 819.829294] env[62600]: DEBUG nova.virt.hardware [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 819.829606] env[62600]: DEBUG nova.virt.hardware [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 819.829995] env[62600]: DEBUG nova.virt.hardware [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 819.830365] env[62600]: DEBUG nova.virt.hardware [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 819.830752] env[62600]: DEBUG nova.virt.hardware [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 819.831893] env[62600]: DEBUG nova.scheduler.client.report [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 819.837546] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f75fee-71c4-48ef-b401-0a322e45c988 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.850949] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba01ccbb-39a6-4ba8-8f61-5c5ff22fd572 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.942222] env[62600]: INFO nova.compute.manager [-] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Took 1.96 seconds to deallocate network for instance. [ 819.947300] env[62600]: DEBUG nova.network.neutron [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Updating instance_info_cache with network_info: [{"id": "a1ade93e-f46e-4261-9a3e-45ffbb2aa092", "address": "fa:16:3e:3b:10:d7", "network": {"id": "f6ef7ce4-a8eb-4d8a-bf1d-cb362e0f442a", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1371992488-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11c88a6316a6409e9ea3d023240c366d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "158692b5-b9fb-49e8-9903-e742ffd6c168", "external-id": "nsx-vlan-transportzone-769", "segmentation_id": 769, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1ade93e-f4", "ovs_interfaceid": "a1ade93e-f46e-4261-9a3e-45ffbb2aa092", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.979426] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f5d346-989d-591a-b684-2ad26c34135e, 'name': SearchDatastore_Task, 'duration_secs': 0.009664} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.979808] env[62600]: DEBUG oslo_concurrency.lockutils [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.980076] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 819.980430] env[62600]: DEBUG oslo_concurrency.lockutils [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.980540] env[62600]: DEBUG oslo_concurrency.lockutils [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.980674] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 819.980938] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6eba1877-bb76-4f7e-a4bb-024b69bac45f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.990950] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 819.991069] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 819.991931] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7dc728e0-db7b-47de-aeab-17fcc2c3798f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.997385] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 819.997385] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5249d993-ba1e-a5cd-2652-5f4029907339" [ 819.997385] env[62600]: _type = "Task" [ 819.997385] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.004901] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5249d993-ba1e-a5cd-2652-5f4029907339, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.057108] env[62600]: DEBUG oslo_vmware.api [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222405, 'name': RemoveSnapshot_Task} progress is 70%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.308060] env[62600]: DEBUG nova.compute.manager [req-724ff8a6-2122-4525-bde4-b6df192138b1 req-81f07018-4fc2-4589-aa88-1844a58cda0c service nova] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Received event network-vif-plugged-6832182f-fe16-4367-a4bc-53d7bd8f6033 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.308060] env[62600]: DEBUG oslo_concurrency.lockutils [req-724ff8a6-2122-4525-bde4-b6df192138b1 req-81f07018-4fc2-4589-aa88-1844a58cda0c service nova] Acquiring lock "dde6df10-618a-40a8-b33f-efc0ca3a9287-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.308060] env[62600]: DEBUG oslo_concurrency.lockutils [req-724ff8a6-2122-4525-bde4-b6df192138b1 req-81f07018-4fc2-4589-aa88-1844a58cda0c service nova] Lock "dde6df10-618a-40a8-b33f-efc0ca3a9287-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.308060] env[62600]: DEBUG oslo_concurrency.lockutils [req-724ff8a6-2122-4525-bde4-b6df192138b1 req-81f07018-4fc2-4589-aa88-1844a58cda0c service nova] Lock "dde6df10-618a-40a8-b33f-efc0ca3a9287-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.308060] env[62600]: DEBUG nova.compute.manager [req-724ff8a6-2122-4525-bde4-b6df192138b1 req-81f07018-4fc2-4589-aa88-1844a58cda0c service nova] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] No waiting events found dispatching network-vif-plugged-6832182f-fe16-4367-a4bc-53d7bd8f6033 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 820.308420] env[62600]: WARNING nova.compute.manager [req-724ff8a6-2122-4525-bde4-b6df192138b1 req-81f07018-4fc2-4589-aa88-1844a58cda0c service nova] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Received unexpected event network-vif-plugged-6832182f-fe16-4367-a4bc-53d7bd8f6033 for instance with vm_state building and task_state spawning. [ 820.345024] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.586s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.345024] env[62600]: DEBUG nova.compute.manager [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 820.348348] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.902s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.350396] env[62600]: INFO nova.compute.claims [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 820.401829] env[62600]: DEBUG nova.network.neutron [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Successfully updated port: 6832182f-fe16-4367-a4bc-53d7bd8f6033 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 820.450274] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Releasing lock "refresh_cache-4e1a376f-6619-4c35-b75f-b45db75815cc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.450651] env[62600]: DEBUG nova.compute.manager [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Instance network_info: |[{"id": "a1ade93e-f46e-4261-9a3e-45ffbb2aa092", "address": "fa:16:3e:3b:10:d7", "network": {"id": "f6ef7ce4-a8eb-4d8a-bf1d-cb362e0f442a", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1371992488-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "11c88a6316a6409e9ea3d023240c366d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "158692b5-b9fb-49e8-9903-e742ffd6c168", "external-id": "nsx-vlan-transportzone-769", "segmentation_id": 769, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1ade93e-f4", "ovs_interfaceid": "a1ade93e-f46e-4261-9a3e-45ffbb2aa092", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 820.451327] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3b:10:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '158692b5-b9fb-49e8-9903-e742ffd6c168', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a1ade93e-f46e-4261-9a3e-45ffbb2aa092', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 820.459899] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Creating folder: Project (11c88a6316a6409e9ea3d023240c366d). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 820.460199] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8e260209-ce9d-460c-8bdc-b95c37e62a89 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.471125] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Created folder: Project (11c88a6316a6409e9ea3d023240c366d) in parent group-v264198. [ 820.471323] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Creating folder: Instances. Parent ref: group-v264260. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 820.471560] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8bef5909-5863-4832-a0dc-cca9de3f3dda {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.480484] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Created folder: Instances in parent group-v264260. [ 820.480728] env[62600]: DEBUG oslo.service.loopingcall [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.480918] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 820.481133] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca495434-8e6c-4467-a231-ab27dd9dae74 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.496654] env[62600]: INFO nova.compute.manager [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Took 0.56 seconds to detach 1 volumes for instance. [ 820.502146] env[62600]: DEBUG nova.compute.manager [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Deleting volume: 88d3934f-d0ae-4578-af24-9abb842074b3 {{(pid=62600) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 820.505436] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 820.505436] env[62600]: value = "task-1222408" [ 820.505436] env[62600]: _type = "Task" [ 820.505436] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.512063] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5249d993-ba1e-a5cd-2652-5f4029907339, 'name': SearchDatastore_Task, 'duration_secs': 0.015186} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.513242] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a6ebc92-f44a-4caf-be50-0bbbab2f1eb2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.519239] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222408, 'name': CreateVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.522553] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 820.522553] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526348f7-6be2-eb71-c826-ad08480dd506" [ 820.522553] env[62600]: _type = "Task" [ 820.522553] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.529899] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526348f7-6be2-eb71-c826-ad08480dd506, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.560859] env[62600]: DEBUG oslo_vmware.api [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222405, 'name': RemoveSnapshot_Task} progress is 98%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.855253] env[62600]: DEBUG nova.compute.utils [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 820.859379] env[62600]: DEBUG nova.compute.manager [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 820.859755] env[62600]: DEBUG nova.network.neutron [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 820.903831] env[62600]: DEBUG oslo_concurrency.lockutils [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquiring lock "refresh_cache-dde6df10-618a-40a8-b33f-efc0ca3a9287" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.903993] env[62600]: DEBUG oslo_concurrency.lockutils [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquired lock "refresh_cache-dde6df10-618a-40a8-b33f-efc0ca3a9287" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.904160] env[62600]: DEBUG nova.network.neutron [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 820.907345] env[62600]: DEBUG nova.policy [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '627da34339ed4ea2a4b0e821e2b9a827', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'db819a7841e345ca8f19b4c11c239f51', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 821.015613] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222408, 'name': CreateVM_Task, 'duration_secs': 0.383508} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.015873] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 821.016558] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.016794] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.017185] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 821.017459] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02bbc91f-d652-4057-8be7-2cf626204f04 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.021936] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Waiting for the task: (returnval){ [ 821.021936] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f77c23-00a9-118b-208a-f2898be2c445" [ 821.021936] env[62600]: _type = "Task" [ 821.021936] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.033665] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f77c23-00a9-118b-208a-f2898be2c445, 'name': SearchDatastore_Task, 'duration_secs': 0.009079} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.036954] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.037261] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 821.037557] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.037861] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526348f7-6be2-eb71-c826-ad08480dd506, 'name': SearchDatastore_Task, 'duration_secs': 0.012325} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.038035] env[62600]: DEBUG oslo_concurrency.lockutils [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.038272] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] bd9b7351-01a2-429d-a860-d9ff6855eefc/bd9b7351-01a2-429d-a860-d9ff6855eefc.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 821.038533] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.038719] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 821.038952] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6b086ab5-0373-4617-9632-52f891bb48a3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.040734] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a93b4a13-f5a5-4bb8-97f5-ffaf2eae1c8b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.046976] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.050430] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 821.050612] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 821.054368] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06693f7d-1ba8-4d77-b127-b6ea3bc30e73 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.056412] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 821.056412] env[62600]: value = "task-1222410" [ 821.056412] env[62600]: _type = "Task" [ 821.056412] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.064674] env[62600]: DEBUG oslo_vmware.api [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222405, 'name': RemoveSnapshot_Task, 'duration_secs': 1.18275} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.064945] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Waiting for the task: (returnval){ [ 821.064945] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5217c789-07a8-b368-66e8-eccc32dce039" [ 821.064945] env[62600]: _type = "Task" [ 821.064945] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.065509] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Deleted Snapshot of the VM instance {{(pid=62600) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 821.065750] env[62600]: INFO nova.compute.manager [None req-9384b373-daea-4dba-8220-8c781b3cdc09 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Took 16.45 seconds to snapshot the instance on the hypervisor. [ 821.073345] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222410, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.079011] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5217c789-07a8-b368-66e8-eccc32dce039, 'name': SearchDatastore_Task, 'duration_secs': 0.009922} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.079752] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c2645ea-5f8b-41b4-967a-a7f5ce3e1968 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.084693] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Waiting for the task: (returnval){ [ 821.084693] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5295cbaa-d4c7-72f1-5059-70e0cbb3cf0c" [ 821.084693] env[62600]: _type = "Task" [ 821.084693] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.092188] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5295cbaa-d4c7-72f1-5059-70e0cbb3cf0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.159188] env[62600]: DEBUG nova.network.neutron [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Successfully created port: 6995d562-b51e-4eb4-8d30-6c89c403144d {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 821.360522] env[62600]: DEBUG nova.compute.manager [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 821.472255] env[62600]: DEBUG nova.network.neutron [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 821.573416] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222410, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.479426} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.574398] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] bd9b7351-01a2-429d-a860-d9ff6855eefc/bd9b7351-01a2-429d-a860-d9ff6855eefc.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 821.574398] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 821.576221] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4e0d293c-720b-40a6-9c87-e9b45a9c03c1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.587034] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 821.587034] env[62600]: value = "task-1222411" [ 821.587034] env[62600]: _type = "Task" [ 821.587034] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.604016] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5295cbaa-d4c7-72f1-5059-70e0cbb3cf0c, 'name': SearchDatastore_Task, 'duration_secs': 0.008616} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.604619] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222411, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.605192] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.605480] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 4e1a376f-6619-4c35-b75f-b45db75815cc/4e1a376f-6619-4c35-b75f-b45db75815cc.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 821.605738] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b5165818-65a1-4d11-a005-546413efcd56 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.616336] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Waiting for the task: (returnval){ [ 821.616336] env[62600]: value = "task-1222412" [ 821.616336] env[62600]: _type = "Task" [ 821.616336] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.626973] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Task: {'id': task-1222412, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.699666] env[62600]: DEBUG nova.network.neutron [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Updating instance_info_cache with network_info: [{"id": "6832182f-fe16-4367-a4bc-53d7bd8f6033", "address": "fa:16:3e:3c:d5:5a", "network": {"id": "1a7b3797-0a3c-4aba-a9dd-6d560f45dc81", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-76817883-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88985cba32da4faaa133d74ffa635292", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd098b1c-636f-492d-b5ae-037cb0cae454", "external-id": "nsx-vlan-transportzone-377", "segmentation_id": 377, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6832182f-fe", "ovs_interfaceid": "6832182f-fe16-4367-a4bc-53d7bd8f6033", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.788112] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90467c10-b443-4ab1-b412-6be46480b9cf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.797156] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86086dc0-1705-4fe6-95d0-81df420228bd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.834797] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abd1371a-63fc-4515-a216-70dceba81a76 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.843650] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0877254c-7928-4ef2-99a0-54545ad6dd81 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.859643] env[62600]: DEBUG nova.compute.provider_tree [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.099695] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222411, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065016} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.099994] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 822.100813] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2dbaeb4-d6aa-4305-8b0b-8d7f96e16810 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.123376] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] bd9b7351-01a2-429d-a860-d9ff6855eefc/bd9b7351-01a2-429d-a860-d9ff6855eefc.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 822.123717] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84aa906f-8541-4a24-9b81-44040b2cc537 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.145677] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Task: {'id': task-1222412, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.42509} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.146855] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 4e1a376f-6619-4c35-b75f-b45db75815cc/4e1a376f-6619-4c35-b75f-b45db75815cc.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 822.147100] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 822.147418] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 822.147418] env[62600]: value = "task-1222413" [ 822.147418] env[62600]: _type = "Task" [ 822.147418] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.147611] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bac9ba63-6212-475c-a14d-95915e89dcd8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.156899] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222413, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.158105] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Waiting for the task: (returnval){ [ 822.158105] env[62600]: value = "task-1222414" [ 822.158105] env[62600]: _type = "Task" [ 822.158105] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.165618] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Task: {'id': task-1222414, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.202207] env[62600]: DEBUG oslo_concurrency.lockutils [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Releasing lock "refresh_cache-dde6df10-618a-40a8-b33f-efc0ca3a9287" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.202690] env[62600]: DEBUG nova.compute.manager [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Instance network_info: |[{"id": "6832182f-fe16-4367-a4bc-53d7bd8f6033", "address": "fa:16:3e:3c:d5:5a", "network": {"id": "1a7b3797-0a3c-4aba-a9dd-6d560f45dc81", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-76817883-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88985cba32da4faaa133d74ffa635292", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd098b1c-636f-492d-b5ae-037cb0cae454", "external-id": "nsx-vlan-transportzone-377", "segmentation_id": 377, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6832182f-fe", "ovs_interfaceid": "6832182f-fe16-4367-a4bc-53d7bd8f6033", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 822.203189] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:d5:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cd098b1c-636f-492d-b5ae-037cb0cae454', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6832182f-fe16-4367-a4bc-53d7bd8f6033', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 822.211691] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Creating folder: Project (88985cba32da4faaa133d74ffa635292). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 822.211769] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-312c0df8-2d27-4c3b-aab0-2bb1afc34ddb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.223564] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Created folder: Project (88985cba32da4faaa133d74ffa635292) in parent group-v264198. [ 822.223786] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Creating folder: Instances. Parent ref: group-v264263. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 822.224063] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a88493cd-3fdb-43b6-9ad9-7ae3b140f158 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.232752] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Created folder: Instances in parent group-v264263. [ 822.232985] env[62600]: DEBUG oslo.service.loopingcall [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.233359] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 822.233581] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8cf9f03d-b22c-483f-a207-421cbdf7f496 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.254156] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 822.254156] env[62600]: value = "task-1222417" [ 822.254156] env[62600]: _type = "Task" [ 822.254156] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.261497] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222417, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.363037] env[62600]: DEBUG nova.scheduler.client.report [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 822.373977] env[62600]: DEBUG nova.compute.manager [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 822.395755] env[62600]: DEBUG nova.virt.hardware [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 822.395869] env[62600]: DEBUG nova.virt.hardware [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 822.396574] env[62600]: DEBUG nova.virt.hardware [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 822.396574] env[62600]: DEBUG nova.virt.hardware [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 822.396881] env[62600]: DEBUG nova.virt.hardware [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 822.396881] env[62600]: DEBUG nova.virt.hardware [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 822.397236] env[62600]: DEBUG nova.virt.hardware [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 822.397236] env[62600]: DEBUG nova.virt.hardware [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 822.400465] env[62600]: DEBUG nova.virt.hardware [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 822.400465] env[62600]: DEBUG nova.virt.hardware [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 822.400465] env[62600]: DEBUG nova.virt.hardware [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 822.400465] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b936b26a-d70c-4206-9d43-676ee9688016 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.408842] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d39732d1-8e31-4a02-8d31-0ddb62b4945f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.458147] env[62600]: DEBUG nova.compute.manager [req-d1fbe640-7a63-4bb1-8323-1582b174206a req-1a4eea52-d413-4c52-9184-36fc06c396cd service nova] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Received event network-changed-6832182f-fe16-4367-a4bc-53d7bd8f6033 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.458147] env[62600]: DEBUG nova.compute.manager [req-d1fbe640-7a63-4bb1-8323-1582b174206a req-1a4eea52-d413-4c52-9184-36fc06c396cd service nova] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Refreshing instance network info cache due to event network-changed-6832182f-fe16-4367-a4bc-53d7bd8f6033. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 822.458147] env[62600]: DEBUG oslo_concurrency.lockutils [req-d1fbe640-7a63-4bb1-8323-1582b174206a req-1a4eea52-d413-4c52-9184-36fc06c396cd service nova] Acquiring lock "refresh_cache-dde6df10-618a-40a8-b33f-efc0ca3a9287" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.458147] env[62600]: DEBUG oslo_concurrency.lockutils [req-d1fbe640-7a63-4bb1-8323-1582b174206a req-1a4eea52-d413-4c52-9184-36fc06c396cd service nova] Acquired lock "refresh_cache-dde6df10-618a-40a8-b33f-efc0ca3a9287" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.458447] env[62600]: DEBUG nova.network.neutron [req-d1fbe640-7a63-4bb1-8323-1582b174206a req-1a4eea52-d413-4c52-9184-36fc06c396cd service nova] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Refreshing network info cache for port 6832182f-fe16-4367-a4bc-53d7bd8f6033 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 822.659973] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222413, 'name': ReconfigVM_Task, 'duration_secs': 0.307823} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.663201] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Reconfigured VM instance instance-00000039 to attach disk [datastore1] bd9b7351-01a2-429d-a860-d9ff6855eefc/bd9b7351-01a2-429d-a860-d9ff6855eefc.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 822.663862] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c8a97f9a-b94d-4b58-be1a-0c86b85a14ac {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.670318] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Task: {'id': task-1222414, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064835} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.671693] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 822.672081] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 822.672081] env[62600]: value = "task-1222418" [ 822.672081] env[62600]: _type = "Task" [ 822.672081] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.672791] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d94513da-dd52-47c4-8a6d-97be0aa17dc7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.684161] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222418, 'name': Rename_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.704659] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] 4e1a376f-6619-4c35-b75f-b45db75815cc/4e1a376f-6619-4c35-b75f-b45db75815cc.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 822.704659] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c980954-f87e-49b2-ae55-7d7d3e78abea {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.724135] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Waiting for the task: (returnval){ [ 822.724135] env[62600]: value = "task-1222419" [ 822.724135] env[62600]: _type = "Task" [ 822.724135] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.733287] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Task: {'id': task-1222419, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.764292] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222417, 'name': CreateVM_Task, 'duration_secs': 0.386188} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.764457] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 822.765226] env[62600]: DEBUG oslo_concurrency.lockutils [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.765525] env[62600]: DEBUG oslo_concurrency.lockutils [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.765825] env[62600]: DEBUG oslo_concurrency.lockutils [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 822.768515] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8abd72e7-223d-4634-8ff5-6d7c5d20f398 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.773889] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for the task: (returnval){ [ 822.773889] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528ae45d-63c7-302c-1237-b2f6638033d4" [ 822.773889] env[62600]: _type = "Task" [ 822.773889] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.781605] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528ae45d-63c7-302c-1237-b2f6638033d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.803759] env[62600]: DEBUG nova.network.neutron [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Successfully updated port: 6995d562-b51e-4eb4-8d30-6c89c403144d {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 822.868579] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.520s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.869546] env[62600]: DEBUG nova.compute.manager [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 822.873905] env[62600]: DEBUG oslo_concurrency.lockutils [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.966s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.874140] env[62600]: DEBUG oslo_concurrency.lockutils [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.876248] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.675s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.877716] env[62600]: INFO nova.compute.claims [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 822.906985] env[62600]: INFO nova.scheduler.client.report [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Deleted allocations for instance dcbe8e2d-6ae9-465f-8394-3978ee61a15b [ 823.156521] env[62600]: DEBUG nova.network.neutron [req-d1fbe640-7a63-4bb1-8323-1582b174206a req-1a4eea52-d413-4c52-9184-36fc06c396cd service nova] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Updated VIF entry in instance network info cache for port 6832182f-fe16-4367-a4bc-53d7bd8f6033. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 823.157044] env[62600]: DEBUG nova.network.neutron [req-d1fbe640-7a63-4bb1-8323-1582b174206a req-1a4eea52-d413-4c52-9184-36fc06c396cd service nova] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Updating instance_info_cache with network_info: [{"id": "6832182f-fe16-4367-a4bc-53d7bd8f6033", "address": "fa:16:3e:3c:d5:5a", "network": {"id": "1a7b3797-0a3c-4aba-a9dd-6d560f45dc81", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-76817883-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88985cba32da4faaa133d74ffa635292", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd098b1c-636f-492d-b5ae-037cb0cae454", "external-id": "nsx-vlan-transportzone-377", "segmentation_id": 377, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6832182f-fe", "ovs_interfaceid": "6832182f-fe16-4367-a4bc-53d7bd8f6033", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.185012] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222418, 'name': Rename_Task, 'duration_secs': 0.14778} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.185305] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 823.185541] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0c30433-098c-4a7c-972c-662ad94f97fc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.191963] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 823.191963] env[62600]: value = "task-1222420" [ 823.191963] env[62600]: _type = "Task" [ 823.191963] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.198494] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222420, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.233429] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Task: {'id': task-1222419, 'name': ReconfigVM_Task, 'duration_secs': 0.28608} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.233726] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Reconfigured VM instance instance-0000003a to attach disk [datastore1] 4e1a376f-6619-4c35-b75f-b45db75815cc/4e1a376f-6619-4c35-b75f-b45db75815cc.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 823.234689] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-76b2c517-cfd4-4336-9fc1-03556a7b6ad6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.240852] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Waiting for the task: (returnval){ [ 823.240852] env[62600]: value = "task-1222421" [ 823.240852] env[62600]: _type = "Task" [ 823.240852] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.248927] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Task: {'id': task-1222421, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.272672] env[62600]: DEBUG nova.compute.manager [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 823.273800] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df819701-3ef5-4f38-bef5-42ae94ff81fc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.289840] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528ae45d-63c7-302c-1237-b2f6638033d4, 'name': SearchDatastore_Task, 'duration_secs': 0.012957} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.290474] env[62600]: DEBUG oslo_concurrency.lockutils [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.290744] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 823.290994] env[62600]: DEBUG oslo_concurrency.lockutils [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.291179] env[62600]: DEBUG oslo_concurrency.lockutils [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.291392] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 823.291678] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97317279-8841-4615-83bf-c7ddd3ea2913 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.300354] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 823.300797] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 823.301387] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-282dcce4-7ec7-49c1-9fd6-bdd8fc18c70f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.306876] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Acquiring lock "refresh_cache-fb6f8e36-8d24-45ea-a6e4-4d768c3b232d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.307050] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Acquired lock "refresh_cache-fb6f8e36-8d24-45ea-a6e4-4d768c3b232d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.307236] env[62600]: DEBUG nova.network.neutron [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 823.308953] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for the task: (returnval){ [ 823.308953] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c03f77-dbb3-52dd-3bd9-72ec6924ace2" [ 823.308953] env[62600]: _type = "Task" [ 823.308953] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.317752] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c03f77-dbb3-52dd-3bd9-72ec6924ace2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.382565] env[62600]: DEBUG nova.compute.utils [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 823.386143] env[62600]: DEBUG nova.compute.manager [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 823.386313] env[62600]: DEBUG nova.network.neutron [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 823.418475] env[62600]: DEBUG oslo_concurrency.lockutils [None req-39da85f5-e366-4fec-8281-93638042e44d tempest-ServerPasswordTestJSON-1600022125 tempest-ServerPasswordTestJSON-1600022125-project-member] Lock "dcbe8e2d-6ae9-465f-8394-3978ee61a15b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.701s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.429929] env[62600]: DEBUG nova.policy [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34319d68370c48ef93fa24a9d9132407', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bab74e3ae78248909dbcd483abbb8da2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 823.660542] env[62600]: DEBUG oslo_concurrency.lockutils [req-d1fbe640-7a63-4bb1-8323-1582b174206a req-1a4eea52-d413-4c52-9184-36fc06c396cd service nova] Releasing lock "refresh_cache-dde6df10-618a-40a8-b33f-efc0ca3a9287" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.702809] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222420, 'name': PowerOnVM_Task} progress is 81%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.754578] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Task: {'id': task-1222421, 'name': Rename_Task, 'duration_secs': 0.138244} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.759123] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 823.759123] env[62600]: DEBUG nova.network.neutron [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Successfully created port: b15ca3aa-0a63-4385-9171-98748d4b988c {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 823.760322] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-028900b1-bcf3-4620-a358-e65fbdd16afa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.770108] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Waiting for the task: (returnval){ [ 823.770108] env[62600]: value = "task-1222422" [ 823.770108] env[62600]: _type = "Task" [ 823.770108] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.782160] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Task: {'id': task-1222422, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.791616] env[62600]: INFO nova.compute.manager [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] instance snapshotting [ 823.794425] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be2f425-1118-4799-8746-78d9e01c686a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.823120] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-454ad13d-ae57-4aed-bbc6-3a48701fe29a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.832937] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c03f77-dbb3-52dd-3bd9-72ec6924ace2, 'name': SearchDatastore_Task, 'duration_secs': 0.009648} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.843248] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc31b01c-0dd5-4728-aeab-3ab8e53f4156 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.847693] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for the task: (returnval){ [ 823.847693] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520c59b9-3b4e-2e95-07aa-2d16f7be688b" [ 823.847693] env[62600]: _type = "Task" [ 823.847693] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.857769] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520c59b9-3b4e-2e95-07aa-2d16f7be688b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.870931] env[62600]: DEBUG nova.network.neutron [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 823.887514] env[62600]: DEBUG nova.compute.manager [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 824.063084] env[62600]: DEBUG nova.network.neutron [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Updating instance_info_cache with network_info: [{"id": "6995d562-b51e-4eb4-8d30-6c89c403144d", "address": "fa:16:3e:a1:df:36", "network": {"id": "10bc9396-85ec-4767-8fe0-12fc7e36b87a", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1857675048-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "db819a7841e345ca8f19b4c11c239f51", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31a7f15-a808-4199-9071-31fd05e316ea", "external-id": "nsx-vlan-transportzone-388", "segmentation_id": 388, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6995d562-b5", "ovs_interfaceid": "6995d562-b51e-4eb4-8d30-6c89c403144d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.204587] env[62600]: DEBUG oslo_vmware.api [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222420, 'name': PowerOnVM_Task, 'duration_secs': 0.847839} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.204936] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 824.205164] env[62600]: INFO nova.compute.manager [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Took 9.40 seconds to spawn the instance on the hypervisor. [ 824.205281] env[62600]: DEBUG nova.compute.manager [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 824.206115] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7709b49a-b4cb-469f-ab65-c097b103a1c3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.281804] env[62600]: DEBUG oslo_vmware.api [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Task: {'id': task-1222422, 'name': PowerOnVM_Task, 'duration_secs': 0.462455} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.282116] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 824.282330] env[62600]: INFO nova.compute.manager [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Took 6.98 seconds to spawn the instance on the hypervisor. [ 824.282513] env[62600]: DEBUG nova.compute.manager [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 824.283314] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017667fb-af66-472d-b22a-39607d2f3672 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.289715] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf9b588-3494-4a99-9f2a-7efa05dce9ea {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.305529] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbaa0f1e-773f-47ca-8a52-60cb27c6d33e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.336899] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f71676b6-47f5-48bd-bfbb-188e3f2b60dd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.345442] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Creating Snapshot of the VM instance {{(pid=62600) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 824.346112] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9a5b0cb2-94af-4bda-ad1a-91e46fa6d751 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.349233] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67463e95-a2dd-4ba4-90e6-fdfbdb6ac500 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.366997] env[62600]: DEBUG nova.compute.provider_tree [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 824.372961] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 824.372961] env[62600]: value = "task-1222423" [ 824.372961] env[62600]: _type = "Task" [ 824.372961] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.373264] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520c59b9-3b4e-2e95-07aa-2d16f7be688b, 'name': SearchDatastore_Task, 'duration_secs': 0.010941} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.374399] env[62600]: DEBUG oslo_concurrency.lockutils [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.374399] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] dde6df10-618a-40a8-b33f-efc0ca3a9287/dde6df10-618a-40a8-b33f-efc0ca3a9287.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 824.377418] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3b66c2d5-885f-4b26-81b4-468df8317520 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.385194] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222423, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.386727] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for the task: (returnval){ [ 824.386727] env[62600]: value = "task-1222424" [ 824.386727] env[62600]: _type = "Task" [ 824.386727] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.402017] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222424, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.538082] env[62600]: DEBUG nova.compute.manager [req-c38ceff4-f86b-4058-8058-3de40cc6c7d9 req-a3fe0ab2-35e7-4adc-894b-05eb6a612acb service nova] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Received event network-vif-plugged-6995d562-b51e-4eb4-8d30-6c89c403144d {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.538323] env[62600]: DEBUG oslo_concurrency.lockutils [req-c38ceff4-f86b-4058-8058-3de40cc6c7d9 req-a3fe0ab2-35e7-4adc-894b-05eb6a612acb service nova] Acquiring lock "fb6f8e36-8d24-45ea-a6e4-4d768c3b232d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.538575] env[62600]: DEBUG oslo_concurrency.lockutils [req-c38ceff4-f86b-4058-8058-3de40cc6c7d9 req-a3fe0ab2-35e7-4adc-894b-05eb6a612acb service nova] Lock "fb6f8e36-8d24-45ea-a6e4-4d768c3b232d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.538755] env[62600]: DEBUG oslo_concurrency.lockutils [req-c38ceff4-f86b-4058-8058-3de40cc6c7d9 req-a3fe0ab2-35e7-4adc-894b-05eb6a612acb service nova] Lock "fb6f8e36-8d24-45ea-a6e4-4d768c3b232d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.538934] env[62600]: DEBUG nova.compute.manager [req-c38ceff4-f86b-4058-8058-3de40cc6c7d9 req-a3fe0ab2-35e7-4adc-894b-05eb6a612acb service nova] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] No waiting events found dispatching network-vif-plugged-6995d562-b51e-4eb4-8d30-6c89c403144d {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 824.539111] env[62600]: WARNING nova.compute.manager [req-c38ceff4-f86b-4058-8058-3de40cc6c7d9 req-a3fe0ab2-35e7-4adc-894b-05eb6a612acb service nova] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Received unexpected event network-vif-plugged-6995d562-b51e-4eb4-8d30-6c89c403144d for instance with vm_state building and task_state spawning. [ 824.539276] env[62600]: DEBUG nova.compute.manager [req-c38ceff4-f86b-4058-8058-3de40cc6c7d9 req-a3fe0ab2-35e7-4adc-894b-05eb6a612acb service nova] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Received event network-changed-6995d562-b51e-4eb4-8d30-6c89c403144d {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.539516] env[62600]: DEBUG nova.compute.manager [req-c38ceff4-f86b-4058-8058-3de40cc6c7d9 req-a3fe0ab2-35e7-4adc-894b-05eb6a612acb service nova] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Refreshing instance network info cache due to event network-changed-6995d562-b51e-4eb4-8d30-6c89c403144d. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 824.539688] env[62600]: DEBUG oslo_concurrency.lockutils [req-c38ceff4-f86b-4058-8058-3de40cc6c7d9 req-a3fe0ab2-35e7-4adc-894b-05eb6a612acb service nova] Acquiring lock "refresh_cache-fb6f8e36-8d24-45ea-a6e4-4d768c3b232d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.564646] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Releasing lock "refresh_cache-fb6f8e36-8d24-45ea-a6e4-4d768c3b232d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.564995] env[62600]: DEBUG nova.compute.manager [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Instance network_info: |[{"id": "6995d562-b51e-4eb4-8d30-6c89c403144d", "address": "fa:16:3e:a1:df:36", "network": {"id": "10bc9396-85ec-4767-8fe0-12fc7e36b87a", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1857675048-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "db819a7841e345ca8f19b4c11c239f51", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31a7f15-a808-4199-9071-31fd05e316ea", "external-id": "nsx-vlan-transportzone-388", "segmentation_id": 388, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6995d562-b5", "ovs_interfaceid": "6995d562-b51e-4eb4-8d30-6c89c403144d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 824.565329] env[62600]: DEBUG oslo_concurrency.lockutils [req-c38ceff4-f86b-4058-8058-3de40cc6c7d9 req-a3fe0ab2-35e7-4adc-894b-05eb6a612acb service nova] Acquired lock "refresh_cache-fb6f8e36-8d24-45ea-a6e4-4d768c3b232d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.565524] env[62600]: DEBUG nova.network.neutron [req-c38ceff4-f86b-4058-8058-3de40cc6c7d9 req-a3fe0ab2-35e7-4adc-894b-05eb6a612acb service nova] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Refreshing network info cache for port 6995d562-b51e-4eb4-8d30-6c89c403144d {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 824.566895] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a1:df:36', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e31a7f15-a808-4199-9071-31fd05e316ea', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6995d562-b51e-4eb4-8d30-6c89c403144d', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 824.575986] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Creating folder: Project (db819a7841e345ca8f19b4c11c239f51). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 824.576944] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-067ff73a-a4ee-4877-8039-bf4b7dcc94f7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.590841] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Created folder: Project (db819a7841e345ca8f19b4c11c239f51) in parent group-v264198. [ 824.590841] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Creating folder: Instances. Parent ref: group-v264266. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 824.590841] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ca538abd-6f96-4fa8-8319-76654a4be205 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.600136] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Created folder: Instances in parent group-v264266. [ 824.600407] env[62600]: DEBUG oslo.service.loopingcall [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 824.600631] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 824.603319] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fec54bf0-216e-452c-8b83-afdccff45041 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.625453] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 824.625453] env[62600]: value = "task-1222427" [ 824.625453] env[62600]: _type = "Task" [ 824.625453] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.634622] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222427, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.726650] env[62600]: INFO nova.compute.manager [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Took 35.91 seconds to build instance. [ 824.809972] env[62600]: INFO nova.compute.manager [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Took 33.78 seconds to build instance. [ 824.891530] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222423, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.903308] env[62600]: DEBUG nova.compute.manager [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 824.906321] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222424, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48773} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.909436] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] dde6df10-618a-40a8-b33f-efc0ca3a9287/dde6df10-618a-40a8-b33f-efc0ca3a9287.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 824.909436] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 824.909436] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-559fe3d1-b2ec-475b-8330-619cfc4b402b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.916276] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for the task: (returnval){ [ 824.916276] env[62600]: value = "task-1222428" [ 824.916276] env[62600]: _type = "Task" [ 824.916276] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.917124] env[62600]: DEBUG nova.scheduler.client.report [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 78 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 824.917503] env[62600]: DEBUG nova.compute.provider_tree [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 78 to 79 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 824.917821] env[62600]: DEBUG nova.compute.provider_tree [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 824.934875] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222428, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.937409] env[62600]: DEBUG nova.virt.hardware [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 824.937695] env[62600]: DEBUG nova.virt.hardware [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 824.937862] env[62600]: DEBUG nova.virt.hardware [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 824.938366] env[62600]: DEBUG nova.virt.hardware [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 824.938366] env[62600]: DEBUG nova.virt.hardware [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 824.938470] env[62600]: DEBUG nova.virt.hardware [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 824.938664] env[62600]: DEBUG nova.virt.hardware [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 824.938848] env[62600]: DEBUG nova.virt.hardware [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 824.939287] env[62600]: DEBUG nova.virt.hardware [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 824.939500] env[62600]: DEBUG nova.virt.hardware [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 824.939788] env[62600]: DEBUG nova.virt.hardware [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 824.941491] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0278d9e2-246e-4e15-80e1-3c5a0a493fe1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.951012] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66272745-1f17-44f7-8c43-8e8625d2b273 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.135502] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222427, 'name': CreateVM_Task, 'duration_secs': 0.369981} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.135862] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 825.136667] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.136978] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.137420] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 825.137875] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcd8b31a-fc7c-412b-b629-555f94e6a598 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.145223] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Waiting for the task: (returnval){ [ 825.145223] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]522b5445-8b67-7025-b82c-5a9a8283e70b" [ 825.145223] env[62600]: _type = "Task" [ 825.145223] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.154212] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]522b5445-8b67-7025-b82c-5a9a8283e70b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.229899] env[62600]: DEBUG oslo_concurrency.lockutils [None req-92a26a06-135a-4a53-8444-075bb5d46324 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "bd9b7351-01a2-429d-a860-d9ff6855eefc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.849s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.308291] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Acquiring lock "e25c631d-3d1b-40d2-9fb1-a65431f991a9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.308589] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Lock "e25c631d-3d1b-40d2-9fb1-a65431f991a9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.308848] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Acquiring lock "e25c631d-3d1b-40d2-9fb1-a65431f991a9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.309059] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Lock "e25c631d-3d1b-40d2-9fb1-a65431f991a9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.309254] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Lock "e25c631d-3d1b-40d2-9fb1-a65431f991a9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.311530] env[62600]: INFO nova.compute.manager [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Terminating instance [ 825.313355] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e098adc4-b7e3-45c9-b09d-86321ae851fd tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Lock "4e1a376f-6619-4c35-b75f-b45db75815cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.680s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.314632] env[62600]: DEBUG nova.compute.manager [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 825.314836] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 825.315675] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b3bc0a-7a7d-4db8-8a49-48db7bd456e1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.324031] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 825.324248] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc03846f-c046-431b-8eb7-80a7932fbcc5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.330995] env[62600]: DEBUG oslo_vmware.api [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Waiting for the task: (returnval){ [ 825.330995] env[62600]: value = "task-1222429" [ 825.330995] env[62600]: _type = "Task" [ 825.330995] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.339983] env[62600]: DEBUG oslo_vmware.api [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': task-1222429, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.390150] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222423, 'name': CreateSnapshot_Task, 'duration_secs': 0.548136} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.390150] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Created Snapshot of the VM instance {{(pid=62600) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 825.390150] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24cec103-7ac5-4b1f-8c72-9dbe43073ebe {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.428575] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.552s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.431887] env[62600]: DEBUG nova.compute.manager [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 825.440239] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 21.933s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.440612] env[62600]: DEBUG nova.objects.instance [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62600) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 825.448482] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222428, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071239} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.448482] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 825.449195] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd06bf3b-70ba-410c-a4f3-2e55bb44f220 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.488035] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] dde6df10-618a-40a8-b33f-efc0ca3a9287/dde6df10-618a-40a8-b33f-efc0ca3a9287.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 825.491994] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd1dd3cb-a0b2-4636-a04c-02953cac0b91 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.525516] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for the task: (returnval){ [ 825.525516] env[62600]: value = "task-1222430" [ 825.525516] env[62600]: _type = "Task" [ 825.525516] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.525516] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222430, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.643668] env[62600]: DEBUG nova.network.neutron [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Successfully updated port: b15ca3aa-0a63-4385-9171-98748d4b988c {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 825.658126] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]522b5445-8b67-7025-b82c-5a9a8283e70b, 'name': SearchDatastore_Task, 'duration_secs': 0.017441} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.659326] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.659326] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 825.659856] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.660153] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.660918] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 825.661679] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-708ce3db-505b-4b1a-83a3-61fd8ebe3da2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.682301] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 825.682552] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 825.688212] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a68d0d7-319b-4dee-8e3c-8d6a52d6c7ac {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.694181] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Waiting for the task: (returnval){ [ 825.694181] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5255bcf2-81b5-225d-762c-a00684867317" [ 825.694181] env[62600]: _type = "Task" [ 825.694181] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.702531] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5255bcf2-81b5-225d-762c-a00684867317, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.731862] env[62600]: DEBUG nova.compute.manager [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 825.841596] env[62600]: DEBUG oslo_vmware.api [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': task-1222429, 'name': PowerOffVM_Task, 'duration_secs': 0.367964} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.842133] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 825.842319] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 825.842821] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-87ef2d9e-c6d3-4783-9b53-4ec8242b2366 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.853064] env[62600]: DEBUG nova.network.neutron [req-c38ceff4-f86b-4058-8058-3de40cc6c7d9 req-a3fe0ab2-35e7-4adc-894b-05eb6a612acb service nova] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Updated VIF entry in instance network info cache for port 6995d562-b51e-4eb4-8d30-6c89c403144d. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 825.853445] env[62600]: DEBUG nova.network.neutron [req-c38ceff4-f86b-4058-8058-3de40cc6c7d9 req-a3fe0ab2-35e7-4adc-894b-05eb6a612acb service nova] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Updating instance_info_cache with network_info: [{"id": "6995d562-b51e-4eb4-8d30-6c89c403144d", "address": "fa:16:3e:a1:df:36", "network": {"id": "10bc9396-85ec-4767-8fe0-12fc7e36b87a", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1857675048-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "db819a7841e345ca8f19b4c11c239f51", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31a7f15-a808-4199-9071-31fd05e316ea", "external-id": "nsx-vlan-transportzone-388", "segmentation_id": 388, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6995d562-b5", "ovs_interfaceid": "6995d562-b51e-4eb4-8d30-6c89c403144d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.911146] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Creating linked-clone VM from snapshot {{(pid=62600) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 825.911578] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6313c7f0-40e4-434d-a2d3-c4b5ffadc7c7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.923210] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 825.923210] env[62600]: value = "task-1222432" [ 825.923210] env[62600]: _type = "Task" [ 825.923210] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.936076] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222432, 'name': CloneVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.937960] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 825.938293] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 825.938570] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Deleting the datastore file [datastore1] e25c631d-3d1b-40d2-9fb1-a65431f991a9 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 825.938908] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9a839069-3c5f-40c0-9818-59bc6bfaf12a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.955659] env[62600]: DEBUG nova.compute.utils [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 825.957854] env[62600]: DEBUG oslo_vmware.api [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Waiting for the task: (returnval){ [ 825.957854] env[62600]: value = "task-1222433" [ 825.957854] env[62600]: _type = "Task" [ 825.957854] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.960464] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8652d18a-7bb3-4026-bca6-103f7b0a5d74 tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.521s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.962508] env[62600]: DEBUG nova.compute.manager [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Not allocating networking since 'none' was specified. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 825.962989] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.804s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.966070] env[62600]: INFO nova.compute.claims [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 825.974274] env[62600]: DEBUG nova.compute.manager [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 825.995976] env[62600]: DEBUG oslo_vmware.api [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': task-1222433, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.032945] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222430, 'name': ReconfigVM_Task, 'duration_secs': 0.310818} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.033719] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Reconfigured VM instance instance-0000003b to attach disk [datastore1] dde6df10-618a-40a8-b33f-efc0ca3a9287/dde6df10-618a-40a8-b33f-efc0ca3a9287.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 826.035093] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a011c6f7-3474-40a0-a5a6-a4f390bc5194 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.047060] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for the task: (returnval){ [ 826.047060] env[62600]: value = "task-1222434" [ 826.047060] env[62600]: _type = "Task" [ 826.047060] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.059697] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222434, 'name': Rename_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.123913] env[62600]: DEBUG nova.compute.manager [req-f5a2017f-ba8b-4fce-8a5a-f03abcc15694 req-093f0e51-b539-4d62-b0d8-80eb87e622fa service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Received event network-changed-75e67d5c-8153-4a74-b875-19d68778a85a {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 826.124144] env[62600]: DEBUG nova.compute.manager [req-f5a2017f-ba8b-4fce-8a5a-f03abcc15694 req-093f0e51-b539-4d62-b0d8-80eb87e622fa service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Refreshing instance network info cache due to event network-changed-75e67d5c-8153-4a74-b875-19d68778a85a. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 826.124373] env[62600]: DEBUG oslo_concurrency.lockutils [req-f5a2017f-ba8b-4fce-8a5a-f03abcc15694 req-093f0e51-b539-4d62-b0d8-80eb87e622fa service nova] Acquiring lock "refresh_cache-bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.124525] env[62600]: DEBUG oslo_concurrency.lockutils [req-f5a2017f-ba8b-4fce-8a5a-f03abcc15694 req-093f0e51-b539-4d62-b0d8-80eb87e622fa service nova] Acquired lock "refresh_cache-bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.124681] env[62600]: DEBUG nova.network.neutron [req-f5a2017f-ba8b-4fce-8a5a-f03abcc15694 req-093f0e51-b539-4d62-b0d8-80eb87e622fa service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Refreshing network info cache for port 75e67d5c-8153-4a74-b875-19d68778a85a {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 826.146779] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "refresh_cache-af7036df-b9f0-4ce6-962a-1edd7c1ea211" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.146922] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired lock "refresh_cache-af7036df-b9f0-4ce6-962a-1edd7c1ea211" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.147096] env[62600]: DEBUG nova.network.neutron [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 826.210043] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5255bcf2-81b5-225d-762c-a00684867317, 'name': SearchDatastore_Task, 'duration_secs': 0.012418} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.210193] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3e92a67-c079-407b-9460-33e99d59cb53 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.215828] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Waiting for the task: (returnval){ [ 826.215828] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f9f0bd-ff54-6d27-e09b-97982edfc570" [ 826.215828] env[62600]: _type = "Task" [ 826.215828] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.226974] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f9f0bd-ff54-6d27-e09b-97982edfc570, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.254870] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.356645] env[62600]: DEBUG oslo_concurrency.lockutils [req-c38ceff4-f86b-4058-8058-3de40cc6c7d9 req-a3fe0ab2-35e7-4adc-894b-05eb6a612acb service nova] Releasing lock "refresh_cache-fb6f8e36-8d24-45ea-a6e4-4d768c3b232d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.435627] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222432, 'name': CloneVM_Task} progress is 94%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.472609] env[62600]: DEBUG oslo_vmware.api [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Task: {'id': task-1222433, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.493524} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.473093] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 826.473434] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 826.473811] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 826.476065] env[62600]: INFO nova.compute.manager [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Took 1.16 seconds to destroy the instance on the hypervisor. [ 826.476065] env[62600]: DEBUG oslo.service.loopingcall [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 826.476065] env[62600]: DEBUG nova.compute.manager [-] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 826.476065] env[62600]: DEBUG nova.network.neutron [-] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 826.561169] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222434, 'name': Rename_Task, 'duration_secs': 0.363185} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.561508] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 826.561779] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a200971b-18e9-4cd0-848f-aa91f3a6e062 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.568533] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for the task: (returnval){ [ 826.568533] env[62600]: value = "task-1222435" [ 826.568533] env[62600]: _type = "Task" [ 826.568533] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.577123] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222435, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.672435] env[62600]: DEBUG nova.compute.manager [req-ba15a561-efdb-4b2b-a9f1-d1876a8b3409 req-46fc3422-4c5f-4b79-bac6-db15982d2d87 service nova] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Received event network-vif-plugged-b15ca3aa-0a63-4385-9171-98748d4b988c {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 826.672673] env[62600]: DEBUG oslo_concurrency.lockutils [req-ba15a561-efdb-4b2b-a9f1-d1876a8b3409 req-46fc3422-4c5f-4b79-bac6-db15982d2d87 service nova] Acquiring lock "af7036df-b9f0-4ce6-962a-1edd7c1ea211-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.672891] env[62600]: DEBUG oslo_concurrency.lockutils [req-ba15a561-efdb-4b2b-a9f1-d1876a8b3409 req-46fc3422-4c5f-4b79-bac6-db15982d2d87 service nova] Lock "af7036df-b9f0-4ce6-962a-1edd7c1ea211-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.673168] env[62600]: DEBUG oslo_concurrency.lockutils [req-ba15a561-efdb-4b2b-a9f1-d1876a8b3409 req-46fc3422-4c5f-4b79-bac6-db15982d2d87 service nova] Lock "af7036df-b9f0-4ce6-962a-1edd7c1ea211-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.673377] env[62600]: DEBUG nova.compute.manager [req-ba15a561-efdb-4b2b-a9f1-d1876a8b3409 req-46fc3422-4c5f-4b79-bac6-db15982d2d87 service nova] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] No waiting events found dispatching network-vif-plugged-b15ca3aa-0a63-4385-9171-98748d4b988c {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 826.673581] env[62600]: WARNING nova.compute.manager [req-ba15a561-efdb-4b2b-a9f1-d1876a8b3409 req-46fc3422-4c5f-4b79-bac6-db15982d2d87 service nova] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Received unexpected event network-vif-plugged-b15ca3aa-0a63-4385-9171-98748d4b988c for instance with vm_state building and task_state spawning. [ 826.673744] env[62600]: DEBUG nova.compute.manager [req-ba15a561-efdb-4b2b-a9f1-d1876a8b3409 req-46fc3422-4c5f-4b79-bac6-db15982d2d87 service nova] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Received event network-changed-b15ca3aa-0a63-4385-9171-98748d4b988c {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 826.673903] env[62600]: DEBUG nova.compute.manager [req-ba15a561-efdb-4b2b-a9f1-d1876a8b3409 req-46fc3422-4c5f-4b79-bac6-db15982d2d87 service nova] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Refreshing instance network info cache due to event network-changed-b15ca3aa-0a63-4385-9171-98748d4b988c. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 826.674093] env[62600]: DEBUG oslo_concurrency.lockutils [req-ba15a561-efdb-4b2b-a9f1-d1876a8b3409 req-46fc3422-4c5f-4b79-bac6-db15982d2d87 service nova] Acquiring lock "refresh_cache-af7036df-b9f0-4ce6-962a-1edd7c1ea211" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.695509] env[62600]: DEBUG nova.network.neutron [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 826.729592] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f9f0bd-ff54-6d27-e09b-97982edfc570, 'name': SearchDatastore_Task, 'duration_secs': 0.021658} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.729941] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.730229] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] fb6f8e36-8d24-45ea-a6e4-4d768c3b232d/fb6f8e36-8d24-45ea-a6e4-4d768c3b232d.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 826.730558] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b931f66e-edb3-4a99-89dc-a83b7f42a29a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.738268] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Waiting for the task: (returnval){ [ 826.738268] env[62600]: value = "task-1222436" [ 826.738268] env[62600]: _type = "Task" [ 826.738268] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.747270] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': task-1222436, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.756324] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Acquiring lock "4e1a376f-6619-4c35-b75f-b45db75815cc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.756631] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Lock "4e1a376f-6619-4c35-b75f-b45db75815cc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.756871] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Acquiring lock "4e1a376f-6619-4c35-b75f-b45db75815cc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.757110] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Lock "4e1a376f-6619-4c35-b75f-b45db75815cc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.757315] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Lock "4e1a376f-6619-4c35-b75f-b45db75815cc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.759819] env[62600]: INFO nova.compute.manager [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Terminating instance [ 826.763264] env[62600]: DEBUG nova.compute.manager [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 826.763507] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 826.765304] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d85191-c6f8-4965-8ffe-803e14ad1a4f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.776461] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 826.783460] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-56c39a33-c140-42dc-b6f8-48db539ea6c6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.797349] env[62600]: DEBUG oslo_vmware.api [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Waiting for the task: (returnval){ [ 826.797349] env[62600]: value = "task-1222437" [ 826.797349] env[62600]: _type = "Task" [ 826.797349] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.817648] env[62600]: DEBUG oslo_vmware.api [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Task: {'id': task-1222437, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.936315] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222432, 'name': CloneVM_Task} progress is 94%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.937571] env[62600]: DEBUG nova.network.neutron [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Updating instance_info_cache with network_info: [{"id": "b15ca3aa-0a63-4385-9171-98748d4b988c", "address": "fa:16:3e:0d:e9:b6", "network": {"id": "5ef67e8a-5613-4cdb-9f35-0a18720657ff", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1644754500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bab74e3ae78248909dbcd483abbb8da2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb15ca3aa-0a", "ovs_interfaceid": "b15ca3aa-0a63-4385-9171-98748d4b988c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.986640] env[62600]: DEBUG nova.compute.manager [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 827.014219] env[62600]: DEBUG nova.virt.hardware [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 827.014526] env[62600]: DEBUG nova.virt.hardware [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 827.014695] env[62600]: DEBUG nova.virt.hardware [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 827.014891] env[62600]: DEBUG nova.virt.hardware [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 827.015084] env[62600]: DEBUG nova.virt.hardware [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 827.015251] env[62600]: DEBUG nova.virt.hardware [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 827.015477] env[62600]: DEBUG nova.virt.hardware [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 827.015648] env[62600]: DEBUG nova.virt.hardware [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 827.015827] env[62600]: DEBUG nova.virt.hardware [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 827.016144] env[62600]: DEBUG nova.virt.hardware [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 827.016204] env[62600]: DEBUG nova.virt.hardware [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 827.017150] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b6231f-f297-4e15-bebe-17b995c5addf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.028027] env[62600]: DEBUG nova.network.neutron [req-f5a2017f-ba8b-4fce-8a5a-f03abcc15694 req-093f0e51-b539-4d62-b0d8-80eb87e622fa service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Updated VIF entry in instance network info cache for port 75e67d5c-8153-4a74-b875-19d68778a85a. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 827.028027] env[62600]: DEBUG nova.network.neutron [req-f5a2017f-ba8b-4fce-8a5a-f03abcc15694 req-093f0e51-b539-4d62-b0d8-80eb87e622fa service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Updating instance_info_cache with network_info: [{"id": "75e67d5c-8153-4a74-b875-19d68778a85a", "address": "fa:16:3e:fa:9d:b3", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75e67d5c-81", "ovs_interfaceid": "75e67d5c-8153-4a74-b875-19d68778a85a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.032572] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e09c00eb-654a-4c4a-b9ad-6f763a7e96eb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.056244] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Instance VIF info [] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 827.059652] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Creating folder: Project (3db1445381734e1b872cff2475977283). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 827.064399] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e1c19363-6b7b-4f16-98c3-ab553c25f1db {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.076201] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Created folder: Project (3db1445381734e1b872cff2475977283) in parent group-v264198. [ 827.076424] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Creating folder: Instances. Parent ref: group-v264271. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 827.077180] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-de25ea2d-a959-4c7f-a225-f4a43ef4ec78 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.084706] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222435, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.095237] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Created folder: Instances in parent group-v264271. [ 827.095522] env[62600]: DEBUG oslo.service.loopingcall [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.095738] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 827.096049] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d3ecee40-bac7-4372-b066-530840ebeb75 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.122904] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 827.122904] env[62600]: value = "task-1222440" [ 827.122904] env[62600]: _type = "Task" [ 827.122904] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.135788] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222440, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.248674] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': task-1222436, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.310926] env[62600]: DEBUG oslo_vmware.api [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Task: {'id': task-1222437, 'name': PowerOffVM_Task, 'duration_secs': 0.179663} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.313601] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 827.313867] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 827.315201] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a31b5624-fb36-415d-a690-7e0182654d3a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.434638] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e916a0-ab40-431a-aee4-24eded0a6fd8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.440386] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Releasing lock "refresh_cache-af7036df-b9f0-4ce6-962a-1edd7c1ea211" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.440749] env[62600]: DEBUG nova.compute.manager [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Instance network_info: |[{"id": "b15ca3aa-0a63-4385-9171-98748d4b988c", "address": "fa:16:3e:0d:e9:b6", "network": {"id": "5ef67e8a-5613-4cdb-9f35-0a18720657ff", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1644754500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bab74e3ae78248909dbcd483abbb8da2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb15ca3aa-0a", "ovs_interfaceid": "b15ca3aa-0a63-4385-9171-98748d4b988c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 827.441070] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222432, 'name': CloneVM_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.441714] env[62600]: DEBUG oslo_concurrency.lockutils [req-ba15a561-efdb-4b2b-a9f1-d1876a8b3409 req-46fc3422-4c5f-4b79-bac6-db15982d2d87 service nova] Acquired lock "refresh_cache-af7036df-b9f0-4ce6-962a-1edd7c1ea211" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.441909] env[62600]: DEBUG nova.network.neutron [req-ba15a561-efdb-4b2b-a9f1-d1876a8b3409 req-46fc3422-4c5f-4b79-bac6-db15982d2d87 service nova] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Refreshing network info cache for port b15ca3aa-0a63-4385-9171-98748d4b988c {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 827.443217] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0d:e9:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3734b156-0f7d-4721-b23c-d000412ec2eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b15ca3aa-0a63-4385-9171-98748d4b988c', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 827.450994] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Creating folder: Project (bab74e3ae78248909dbcd483abbb8da2). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 827.454369] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-737d0b00-d55e-42f3-ae9d-a0e12aac7ad6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.457156] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b949edbb-cdb2-4e4c-99ec-14e8e874d015 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.492241] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ed0f8f6-26a1-495f-bc69-e7cfacab76ea {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.493480] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Created folder: Project (bab74e3ae78248909dbcd483abbb8da2) in parent group-v264198. [ 827.493655] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Creating folder: Instances. Parent ref: group-v264274. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 827.493897] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4b7bd4be-4035-46dc-a19e-5bb09255c3d5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.500599] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab468c1f-96dd-4d23-a122-d3f60374e9da {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.506194] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Created folder: Instances in parent group-v264274. [ 827.506441] env[62600]: DEBUG oslo.service.loopingcall [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.506633] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 827.506829] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5685fa38-57eb-44bc-9c1c-036415cf507a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.532327] env[62600]: DEBUG oslo_concurrency.lockutils [req-f5a2017f-ba8b-4fce-8a5a-f03abcc15694 req-093f0e51-b539-4d62-b0d8-80eb87e622fa service nova] Releasing lock "refresh_cache-bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.533077] env[62600]: DEBUG nova.compute.provider_tree [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 827.535723] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 827.535804] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 827.535969] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Deleting the datastore file [datastore1] 4e1a376f-6619-4c35-b75f-b45db75815cc {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 827.536432] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f3a484f0-77ca-445b-9435-6cb7ba3df426 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.539538] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 827.539538] env[62600]: value = "task-1222444" [ 827.539538] env[62600]: _type = "Task" [ 827.539538] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.544165] env[62600]: DEBUG oslo_vmware.api [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Waiting for the task: (returnval){ [ 827.544165] env[62600]: value = "task-1222445" [ 827.544165] env[62600]: _type = "Task" [ 827.544165] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.551019] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222444, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.555745] env[62600]: DEBUG oslo_vmware.api [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Task: {'id': task-1222445, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.581400] env[62600]: DEBUG oslo_vmware.api [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222435, 'name': PowerOnVM_Task, 'duration_secs': 0.549766} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.581642] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 827.581812] env[62600]: INFO nova.compute.manager [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Took 7.78 seconds to spawn the instance on the hypervisor. [ 827.582034] env[62600]: DEBUG nova.compute.manager [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 827.582924] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4f63b86-a353-4e3c-ad0a-d4c6b4f722e7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.634095] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222440, 'name': CreateVM_Task, 'duration_secs': 0.294587} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.634303] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 827.634949] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.635394] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.635394] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 827.635897] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01035302-422f-4e34-9f0e-9bb988016926 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.641526] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for the task: (returnval){ [ 827.641526] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]527ddf5a-0d52-cc6d-d5fb-8e065bd3d706" [ 827.641526] env[62600]: _type = "Task" [ 827.641526] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.650374] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]527ddf5a-0d52-cc6d-d5fb-8e065bd3d706, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.749652] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': task-1222436, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.869112] env[62600]: DEBUG nova.network.neutron [-] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.939709] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222432, 'name': CloneVM_Task, 'duration_secs': 1.579368} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.939998] env[62600]: INFO nova.virt.vmwareapi.vmops [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Created linked-clone VM from snapshot [ 827.941484] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fae9160-b18c-4e65-810d-f42565aef67e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.950212] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Uploading image 2685df45-5a39-4d23-9105-ef28747ec7d4 {{(pid=62600) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 827.979862] env[62600]: DEBUG oslo_vmware.rw_handles [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 827.979862] env[62600]: value = "vm-264270" [ 827.979862] env[62600]: _type = "VirtualMachine" [ 827.979862] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 827.979862] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0e8315c2-abba-4fa8-a54b-99905616f485 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.985265] env[62600]: DEBUG oslo_vmware.rw_handles [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lease: (returnval){ [ 827.985265] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5231bf5b-b7aa-cf8d-125b-7c797649272f" [ 827.985265] env[62600]: _type = "HttpNfcLease" [ 827.985265] env[62600]: } obtained for exporting VM: (result){ [ 827.985265] env[62600]: value = "vm-264270" [ 827.985265] env[62600]: _type = "VirtualMachine" [ 827.985265] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 827.985487] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the lease: (returnval){ [ 827.985487] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5231bf5b-b7aa-cf8d-125b-7c797649272f" [ 827.985487] env[62600]: _type = "HttpNfcLease" [ 827.985487] env[62600]: } to be ready. {{(pid=62600) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 827.992496] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 827.992496] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5231bf5b-b7aa-cf8d-125b-7c797649272f" [ 827.992496] env[62600]: _type = "HttpNfcLease" [ 827.992496] env[62600]: } is initializing. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 828.052189] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222444, 'name': CreateVM_Task} progress is 25%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.057614] env[62600]: DEBUG oslo_vmware.api [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Task: {'id': task-1222445, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.075585] env[62600]: DEBUG nova.scheduler.client.report [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 79 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 828.075893] env[62600]: DEBUG nova.compute.provider_tree [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 79 to 80 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 828.076109] env[62600]: DEBUG nova.compute.provider_tree [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 828.109633] env[62600]: INFO nova.compute.manager [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Took 33.74 seconds to build instance. [ 828.155525] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]527ddf5a-0d52-cc6d-d5fb-8e065bd3d706, 'name': SearchDatastore_Task, 'duration_secs': 0.0122} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.155681] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.155926] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 828.156188] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.156343] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.156529] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 828.157678] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aa638c24-57b5-418e-8296-9ba2e4803f03 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.167989] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 828.168213] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 828.168954] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa8c056a-3e76-493c-a56b-98b7fc8c4abc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.174207] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for the task: (returnval){ [ 828.174207] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5242cf29-157e-fe54-a629-832662178306" [ 828.174207] env[62600]: _type = "Task" [ 828.174207] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.186817] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5242cf29-157e-fe54-a629-832662178306, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.250179] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': task-1222436, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.311542] env[62600]: DEBUG nova.compute.manager [req-5f4bd702-8efb-4033-93b3-91b204abc9f6 req-3771654c-dc65-4c04-846f-308745df1560 service nova] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Received event network-vif-deleted-8ca7491d-af98-4a32-b8d9-63e6b57356ef {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 828.321288] env[62600]: DEBUG nova.network.neutron [req-ba15a561-efdb-4b2b-a9f1-d1876a8b3409 req-46fc3422-4c5f-4b79-bac6-db15982d2d87 service nova] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Updated VIF entry in instance network info cache for port b15ca3aa-0a63-4385-9171-98748d4b988c. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 828.321288] env[62600]: DEBUG nova.network.neutron [req-ba15a561-efdb-4b2b-a9f1-d1876a8b3409 req-46fc3422-4c5f-4b79-bac6-db15982d2d87 service nova] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Updating instance_info_cache with network_info: [{"id": "b15ca3aa-0a63-4385-9171-98748d4b988c", "address": "fa:16:3e:0d:e9:b6", "network": {"id": "5ef67e8a-5613-4cdb-9f35-0a18720657ff", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1644754500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bab74e3ae78248909dbcd483abbb8da2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb15ca3aa-0a", "ovs_interfaceid": "b15ca3aa-0a63-4385-9171-98748d4b988c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.373030] env[62600]: INFO nova.compute.manager [-] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Took 1.90 seconds to deallocate network for instance. [ 828.433576] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Acquiring lock "0a8840f1-7681-4aaa-9f3a-32b72c04c1c4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.433913] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Lock "0a8840f1-7681-4aaa-9f3a-32b72c04c1c4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.434180] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Acquiring lock "0a8840f1-7681-4aaa-9f3a-32b72c04c1c4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.434388] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Lock "0a8840f1-7681-4aaa-9f3a-32b72c04c1c4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.434568] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Lock "0a8840f1-7681-4aaa-9f3a-32b72c04c1c4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.436695] env[62600]: INFO nova.compute.manager [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Terminating instance [ 828.439040] env[62600]: DEBUG nova.compute.manager [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 828.439252] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 828.440258] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fd5e50e-210d-4275-9377-ac1d8959b6fd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.448090] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 828.448350] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ff60c13-4e84-410f-b037-c0884d1b275e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.454382] env[62600]: DEBUG oslo_vmware.api [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Waiting for the task: (returnval){ [ 828.454382] env[62600]: value = "task-1222447" [ 828.454382] env[62600]: _type = "Task" [ 828.454382] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.462834] env[62600]: DEBUG oslo_vmware.api [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': task-1222447, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.495957] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 828.495957] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5231bf5b-b7aa-cf8d-125b-7c797649272f" [ 828.495957] env[62600]: _type = "HttpNfcLease" [ 828.495957] env[62600]: } is ready. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 828.496299] env[62600]: DEBUG oslo_vmware.rw_handles [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 828.496299] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5231bf5b-b7aa-cf8d-125b-7c797649272f" [ 828.496299] env[62600]: _type = "HttpNfcLease" [ 828.496299] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 828.497097] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f168f23e-c08c-4269-912d-a83ddc36fa7f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.506733] env[62600]: DEBUG oslo_vmware.rw_handles [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5244e429-4da1-4429-476a-de6d515f9670/disk-0.vmdk from lease info. {{(pid=62600) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 828.506858] env[62600]: DEBUG oslo_vmware.rw_handles [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5244e429-4da1-4429-476a-de6d515f9670/disk-0.vmdk for reading. {{(pid=62600) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 828.578829] env[62600]: DEBUG oslo_vmware.api [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Task: {'id': task-1222445, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.604393} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.579777] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 828.580417] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 828.580696] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 828.580911] env[62600]: INFO nova.compute.manager [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Took 1.82 seconds to destroy the instance on the hypervisor. [ 828.581172] env[62600]: DEBUG oslo.service.loopingcall [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.581352] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222444, 'name': CreateVM_Task, 'duration_secs': 0.710647} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.582081] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.619s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.582889] env[62600]: DEBUG nova.compute.manager [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 828.585133] env[62600]: DEBUG nova.compute.manager [-] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 828.585234] env[62600]: DEBUG nova.network.neutron [-] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 828.586792] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 828.587332] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.618s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.587572] env[62600]: DEBUG nova.objects.instance [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lazy-loading 'resources' on Instance uuid d34c098d-a827-4dbd-96e4-e27d9d56b847 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 828.588991] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.589169] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.589487] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 828.590240] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6344138f-b695-4f74-9c98-aa4cb4e8c797 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.595770] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 828.595770] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52affcb7-40aa-c955-061b-d0e2a55b91dc" [ 828.595770] env[62600]: _type = "Task" [ 828.595770] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.604882] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52affcb7-40aa-c955-061b-d0e2a55b91dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.614838] env[62600]: DEBUG oslo_concurrency.lockutils [None req-963df13d-8143-4eb0-848c-d400f89b072f tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.254s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.641249] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3c3bee56-a9d2-439d-a237-888477206232 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.692950] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5242cf29-157e-fe54-a629-832662178306, 'name': SearchDatastore_Task, 'duration_secs': 0.022827} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.692950] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95eb5c4b-6d41-4474-9a7e-3a715d37cbb6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.700019] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for the task: (returnval){ [ 828.700019] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]527f128a-0645-4812-cca6-284f75dad3e0" [ 828.700019] env[62600]: _type = "Task" [ 828.700019] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.706361] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]527f128a-0645-4812-cca6-284f75dad3e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.754020] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': task-1222436, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.560161} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.754020] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] fb6f8e36-8d24-45ea-a6e4-4d768c3b232d/fb6f8e36-8d24-45ea-a6e4-4d768c3b232d.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 828.754020] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.754020] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f52012a9-c22b-48ac-9e02-7bdcb6d976c5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.762748] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Waiting for the task: (returnval){ [ 828.762748] env[62600]: value = "task-1222448" [ 828.762748] env[62600]: _type = "Task" [ 828.762748] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.771671] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': task-1222448, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.824361] env[62600]: DEBUG oslo_concurrency.lockutils [req-ba15a561-efdb-4b2b-a9f1-d1876a8b3409 req-46fc3422-4c5f-4b79-bac6-db15982d2d87 service nova] Releasing lock "refresh_cache-af7036df-b9f0-4ce6-962a-1edd7c1ea211" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.877841] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.965083] env[62600]: DEBUG oslo_vmware.api [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': task-1222447, 'name': PowerOffVM_Task, 'duration_secs': 0.268258} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.965704] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 828.966043] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 828.967161] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b3b8b5e0-cc5b-44ad-b6cd-4edeb9f00a5f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.994099] env[62600]: DEBUG nova.compute.manager [req-4b67a5ef-3b96-48a4-b329-007a889ef67e req-a7fa883a-6de8-4dd7-8d11-9bc4af4b038f service nova] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Received event network-changed-6832182f-fe16-4367-a4bc-53d7bd8f6033 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 828.994316] env[62600]: DEBUG nova.compute.manager [req-4b67a5ef-3b96-48a4-b329-007a889ef67e req-a7fa883a-6de8-4dd7-8d11-9bc4af4b038f service nova] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Refreshing instance network info cache due to event network-changed-6832182f-fe16-4367-a4bc-53d7bd8f6033. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 828.994533] env[62600]: DEBUG oslo_concurrency.lockutils [req-4b67a5ef-3b96-48a4-b329-007a889ef67e req-a7fa883a-6de8-4dd7-8d11-9bc4af4b038f service nova] Acquiring lock "refresh_cache-dde6df10-618a-40a8-b33f-efc0ca3a9287" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.994676] env[62600]: DEBUG oslo_concurrency.lockutils [req-4b67a5ef-3b96-48a4-b329-007a889ef67e req-a7fa883a-6de8-4dd7-8d11-9bc4af4b038f service nova] Acquired lock "refresh_cache-dde6df10-618a-40a8-b33f-efc0ca3a9287" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.994840] env[62600]: DEBUG nova.network.neutron [req-4b67a5ef-3b96-48a4-b329-007a889ef67e req-a7fa883a-6de8-4dd7-8d11-9bc4af4b038f service nova] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Refreshing network info cache for port 6832182f-fe16-4367-a4bc-53d7bd8f6033 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 829.032713] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 829.033171] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 829.033465] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Deleting the datastore file [datastore2] 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 829.033807] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e9622821-1cc1-4f88-9495-934215db6c67 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.043516] env[62600]: DEBUG oslo_vmware.api [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Waiting for the task: (returnval){ [ 829.043516] env[62600]: value = "task-1222450" [ 829.043516] env[62600]: _type = "Task" [ 829.043516] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.057402] env[62600]: DEBUG oslo_vmware.api [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': task-1222450, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.089333] env[62600]: DEBUG nova.compute.utils [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 829.091151] env[62600]: DEBUG nova.compute.manager [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 829.091387] env[62600]: DEBUG nova.network.neutron [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 829.107328] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52affcb7-40aa-c955-061b-d0e2a55b91dc, 'name': SearchDatastore_Task, 'duration_secs': 0.009913} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.108720] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.109170] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 829.109827] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.149239] env[62600]: DEBUG nova.policy [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '46e18992b34d412e8e3e8205d1d18265', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '734eda84fbe4493ba1f3db243a2bdb09', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 829.212956] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]527f128a-0645-4812-cca6-284f75dad3e0, 'name': SearchDatastore_Task, 'duration_secs': 0.022973} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.213461] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.213830] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] c2c20e5f-6c0e-4989-bc36-ff4a27d5c534/c2c20e5f-6c0e-4989-bc36-ff4a27d5c534.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 829.214166] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.214537] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 829.214792] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-76316a7f-f16b-4cc1-aaf5-372f917c6457 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.220619] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d282a447-1e07-454a-9599-9aff3799273b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.228648] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for the task: (returnval){ [ 829.228648] env[62600]: value = "task-1222451" [ 829.228648] env[62600]: _type = "Task" [ 829.228648] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.231747] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 829.231747] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 829.234919] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c1f92b4-8dc0-4d63-99d7-ea662b491c4f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.242464] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222451, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.246355] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 829.246355] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]522ba5cc-724e-1c41-030f-ddfb0119d908" [ 829.246355] env[62600]: _type = "Task" [ 829.246355] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.256449] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]522ba5cc-724e-1c41-030f-ddfb0119d908, 'name': SearchDatastore_Task, 'duration_secs': 0.011078} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.257786] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-458112ab-c5f8-4cbb-a3c1-d81dc1969b2e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.268988] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 829.268988] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52a75e61-451a-efd9-a93e-98aba6218c35" [ 829.268988] env[62600]: _type = "Task" [ 829.268988] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.276077] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': task-1222448, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057408} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.279558] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 829.280659] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c753bf35-0d5b-48a0-a115-b59c3b705eb9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.287721] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52a75e61-451a-efd9-a93e-98aba6218c35, 'name': SearchDatastore_Task, 'duration_secs': 0.009552} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.288542] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.288987] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] af7036df-b9f0-4ce6-962a-1edd7c1ea211/af7036df-b9f0-4ce6-962a-1edd7c1ea211.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 829.289386] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bb5edc3f-f20c-4587-b65a-547cd3c6e95c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.311709] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] fb6f8e36-8d24-45ea-a6e4-4d768c3b232d/fb6f8e36-8d24-45ea-a6e4-4d768c3b232d.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 829.315550] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-86c72e70-426a-4761-b9dc-8011c9f8fb5b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.333706] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 829.333706] env[62600]: value = "task-1222452" [ 829.333706] env[62600]: _type = "Task" [ 829.333706] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.338766] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Waiting for the task: (returnval){ [ 829.338766] env[62600]: value = "task-1222453" [ 829.338766] env[62600]: _type = "Task" [ 829.338766] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.347709] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222452, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.351597] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': task-1222453, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.520926] env[62600]: DEBUG nova.network.neutron [-] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.561045] env[62600]: DEBUG oslo_vmware.api [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Task: {'id': task-1222450, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.263793} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.561415] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 829.561659] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 829.561659] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 829.561969] env[62600]: INFO nova.compute.manager [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Took 1.12 seconds to destroy the instance on the hypervisor. [ 829.562292] env[62600]: DEBUG oslo.service.loopingcall [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.565793] env[62600]: DEBUG nova.compute.manager [-] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 829.566041] env[62600]: DEBUG nova.network.neutron [-] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 829.600792] env[62600]: DEBUG nova.compute.manager [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 829.606739] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aef6bff-5a6a-43e6-b769-4d87f5085a5a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.619357] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-973904c6-2f50-4e8f-af85-9d403c10b4dd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.662680] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b47f854f-33c1-4db6-b043-ee03a9a36b24 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.668611] env[62600]: DEBUG nova.network.neutron [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Successfully created port: a3ce62f1-b1fb-40f1-a551-e9d26d299144 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 829.681308] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1109e482-0a5c-4103-b2a8-f3af400dc764 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.698514] env[62600]: DEBUG nova.compute.provider_tree [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.742233] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222451, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.848779] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222452, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.854992] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': task-1222453, 'name': ReconfigVM_Task, 'duration_secs': 0.36291} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.855868] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Reconfigured VM instance instance-0000003c to attach disk [datastore1] fb6f8e36-8d24-45ea-a6e4-4d768c3b232d/fb6f8e36-8d24-45ea-a6e4-4d768c3b232d.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 829.857072] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-17ff8fe2-96be-4294-a611-c8f3161aaa4e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.864210] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Waiting for the task: (returnval){ [ 829.864210] env[62600]: value = "task-1222454" [ 829.864210] env[62600]: _type = "Task" [ 829.864210] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.875255] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': task-1222454, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.031625] env[62600]: INFO nova.compute.manager [-] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Took 1.45 seconds to deallocate network for instance. [ 830.095541] env[62600]: DEBUG nova.network.neutron [req-4b67a5ef-3b96-48a4-b329-007a889ef67e req-a7fa883a-6de8-4dd7-8d11-9bc4af4b038f service nova] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Updated VIF entry in instance network info cache for port 6832182f-fe16-4367-a4bc-53d7bd8f6033. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 830.095954] env[62600]: DEBUG nova.network.neutron [req-4b67a5ef-3b96-48a4-b329-007a889ef67e req-a7fa883a-6de8-4dd7-8d11-9bc4af4b038f service nova] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Updating instance_info_cache with network_info: [{"id": "6832182f-fe16-4367-a4bc-53d7bd8f6033", "address": "fa:16:3e:3c:d5:5a", "network": {"id": "1a7b3797-0a3c-4aba-a9dd-6d560f45dc81", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-76817883-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "88985cba32da4faaa133d74ffa635292", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd098b1c-636f-492d-b5ae-037cb0cae454", "external-id": "nsx-vlan-transportzone-377", "segmentation_id": 377, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6832182f-fe", "ovs_interfaceid": "6832182f-fe16-4367-a4bc-53d7bd8f6033", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.202937] env[62600]: DEBUG nova.scheduler.client.report [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 830.242068] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222451, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518359} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.242378] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] c2c20e5f-6c0e-4989-bc36-ff4a27d5c534/c2c20e5f-6c0e-4989-bc36-ff4a27d5c534.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 830.242681] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 830.242976] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cda9bac6-97e9-4ca5-9554-9153b000e6ec {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.250166] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for the task: (returnval){ [ 830.250166] env[62600]: value = "task-1222455" [ 830.250166] env[62600]: _type = "Task" [ 830.250166] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.258334] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222455, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.345664] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222452, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.76139} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.345980] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] af7036df-b9f0-4ce6-962a-1edd7c1ea211/af7036df-b9f0-4ce6-962a-1edd7c1ea211.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 830.346205] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 830.346554] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-66f0f854-c613-4aa2-bc67-3d30381e8a80 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.355078] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 830.355078] env[62600]: value = "task-1222456" [ 830.355078] env[62600]: _type = "Task" [ 830.355078] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.364922] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222456, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.374238] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': task-1222454, 'name': Rename_Task, 'duration_secs': 0.150019} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.375498] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 830.376739] env[62600]: DEBUG nova.compute.manager [req-0af5cf8c-53ec-4dae-9b34-5ef1014e139d req-2fd49e5d-fdfb-43f7-833d-c93b97b9ccdf service nova] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Received event network-vif-deleted-a1ade93e-f46e-4261-9a3e-45ffbb2aa092 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 830.377083] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ed245955-1911-45f8-9df4-263fd333095f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.385028] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Waiting for the task: (returnval){ [ 830.385028] env[62600]: value = "task-1222457" [ 830.385028] env[62600]: _type = "Task" [ 830.385028] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.394615] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': task-1222457, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.542423] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.600116] env[62600]: DEBUG oslo_concurrency.lockutils [req-4b67a5ef-3b96-48a4-b329-007a889ef67e req-a7fa883a-6de8-4dd7-8d11-9bc4af4b038f service nova] Releasing lock "refresh_cache-dde6df10-618a-40a8-b33f-efc0ca3a9287" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.611331] env[62600]: DEBUG nova.compute.manager [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 830.654341] env[62600]: DEBUG nova.virt.hardware [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 830.655250] env[62600]: DEBUG nova.virt.hardware [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 830.655645] env[62600]: DEBUG nova.virt.hardware [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 830.655976] env[62600]: DEBUG nova.virt.hardware [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 830.656289] env[62600]: DEBUG nova.virt.hardware [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 830.657145] env[62600]: DEBUG nova.virt.hardware [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 830.657357] env[62600]: DEBUG nova.virt.hardware [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 830.657588] env[62600]: DEBUG nova.virt.hardware [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 830.657868] env[62600]: DEBUG nova.virt.hardware [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 830.658194] env[62600]: DEBUG nova.virt.hardware [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 830.658475] env[62600]: DEBUG nova.virt.hardware [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 830.659535] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a5c9de4-0d1a-4409-964c-9fd40ecc1d90 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.668821] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e5bf3ca-349b-43bc-9ed0-4bf3fc556e15 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.709397] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.122s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.712098] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.460s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.713653] env[62600]: INFO nova.compute.claims [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 830.732128] env[62600]: INFO nova.scheduler.client.report [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Deleted allocations for instance d34c098d-a827-4dbd-96e4-e27d9d56b847 [ 830.760772] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222455, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076219} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.762171] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 830.763227] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dfcd199-ce98-49a7-9d46-f76f40982897 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.784925] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] c2c20e5f-6c0e-4989-bc36-ff4a27d5c534/c2c20e5f-6c0e-4989-bc36-ff4a27d5c534.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 830.785620] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8cb39c1-ba53-4b46-bb30-9560ba9187d5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.807037] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for the task: (returnval){ [ 830.807037] env[62600]: value = "task-1222458" [ 830.807037] env[62600]: _type = "Task" [ 830.807037] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.816230] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222458, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.852831] env[62600]: DEBUG nova.network.neutron [-] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.866565] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222456, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069737} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.867966] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 830.868297] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71bba9a4-5de1-49ba-8bcf-e27b3a2c1ea5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.891873] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] af7036df-b9f0-4ce6-962a-1edd7c1ea211/af7036df-b9f0-4ce6-962a-1edd7c1ea211.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 830.892441] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e2f57a7-7587-4ebb-a6b4-90393449d826 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.916448] env[62600]: DEBUG oslo_vmware.api [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': task-1222457, 'name': PowerOnVM_Task, 'duration_secs': 0.490153} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.917850] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 830.918199] env[62600]: INFO nova.compute.manager [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Took 8.54 seconds to spawn the instance on the hypervisor. [ 830.918404] env[62600]: DEBUG nova.compute.manager [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 830.918747] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 830.918747] env[62600]: value = "task-1222459" [ 830.918747] env[62600]: _type = "Task" [ 830.918747] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.919453] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a821e96-fe5a-43cb-8b80-27cbcbfc0438 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.930432] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222459, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.023752] env[62600]: DEBUG nova.compute.manager [req-b0e1fce2-da61-46fb-94ac-6643b9d184d1 req-129d569c-b11a-4504-bc37-a68cdcbd0fca service nova] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Received event network-vif-deleted-194093a1-33d0-42f8-8255-db5802300692 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.239970] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4673f5b5-5307-417f-aa4e-ee9b477d182e tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "d34c098d-a827-4dbd-96e4-e27d9d56b847" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.972s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.319122] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222458, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.355366] env[62600]: INFO nova.compute.manager [-] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Took 1.79 seconds to deallocate network for instance. [ 831.432717] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222459, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.445042] env[62600]: INFO nova.compute.manager [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Took 36.39 seconds to build instance. [ 831.491533] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquiring lock "b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.491894] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.492248] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquiring lock "b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.492489] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.492691] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.495178] env[62600]: INFO nova.compute.manager [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Terminating instance [ 831.499675] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquiring lock "refresh_cache-b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.499675] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquired lock "refresh_cache-b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.499675] env[62600]: DEBUG nova.network.neutron [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 831.820564] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222458, 'name': ReconfigVM_Task, 'duration_secs': 0.652641} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.821570] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Reconfigured VM instance instance-0000003e to attach disk [datastore2] c2c20e5f-6c0e-4989-bc36-ff4a27d5c534/c2c20e5f-6c0e-4989-bc36-ff4a27d5c534.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 831.822804] env[62600]: DEBUG nova.network.neutron [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Successfully updated port: a3ce62f1-b1fb-40f1-a551-e9d26d299144 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 831.824275] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e11ee2d0-c049-4054-9a50-d6b479193772 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.831310] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for the task: (returnval){ [ 831.831310] env[62600]: value = "task-1222460" [ 831.831310] env[62600]: _type = "Task" [ 831.831310] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.843874] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222460, 'name': Rename_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.863709] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.932148] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222459, 'name': ReconfigVM_Task, 'duration_secs': 0.601766} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.934893] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Reconfigured VM instance instance-0000003d to attach disk [datastore2] af7036df-b9f0-4ce6-962a-1edd7c1ea211/af7036df-b9f0-4ce6-962a-1edd7c1ea211.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 831.936016] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7c6098e6-8829-4f24-977f-b0f4b0585c74 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.942427] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 831.942427] env[62600]: value = "task-1222461" [ 831.942427] env[62600]: _type = "Task" [ 831.942427] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.947138] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa18ff0e-c725-48ba-80e3-cb0f92ebbfa5 tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Lock "fb6f8e36-8d24-45ea-a6e4-4d768c3b232d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.907s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.952736] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222461, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.031331] env[62600]: DEBUG nova.network.neutron [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 832.067397] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f5dc3b-8562-43d5-a949-6b248ffd03a5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.075660] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd6708c-aeff-4bea-81fb-ed66c6a197d8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.109809] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12b0f395-b758-48a9-a371-1830f51eb0f7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.113333] env[62600]: DEBUG nova.network.neutron [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.120467] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1601db2-cbe3-4799-822a-1336d8c57cb6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.136322] env[62600]: DEBUG nova.compute.provider_tree [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 832.329178] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "refresh_cache-9d899d96-9f4f-41d1-a368-3fde5efc110a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.329178] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquired lock "refresh_cache-9d899d96-9f4f-41d1-a368-3fde5efc110a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.329178] env[62600]: DEBUG nova.network.neutron [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 832.341653] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222460, 'name': Rename_Task, 'duration_secs': 0.189357} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.342294] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 832.342712] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-010efe61-59f8-4e6b-a90f-c22447678415 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.350776] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for the task: (returnval){ [ 832.350776] env[62600]: value = "task-1222462" [ 832.350776] env[62600]: _type = "Task" [ 832.350776] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.360289] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222462, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.454541] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222461, 'name': Rename_Task, 'duration_secs': 0.176243} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.454541] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 832.454541] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1f124393-2352-48cd-9162-12c00f01bed4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.461036] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 832.461036] env[62600]: value = "task-1222463" [ 832.461036] env[62600]: _type = "Task" [ 832.461036] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.467864] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222463, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.615997] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Releasing lock "refresh_cache-b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.616736] env[62600]: DEBUG nova.compute.manager [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 832.617049] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 832.618168] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c60557bf-6670-403e-9f50-0914c41eb366 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.626229] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 832.626529] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e542650e-0af0-448f-9f14-05ceb718186f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.632653] env[62600]: DEBUG oslo_vmware.api [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 832.632653] env[62600]: value = "task-1222464" [ 832.632653] env[62600]: _type = "Task" [ 832.632653] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.638898] env[62600]: DEBUG nova.compute.manager [req-90bd954c-40c5-461e-b59b-6d1ac311527a req-7c781b2d-7699-48d3-b658-e1cb91864e12 service nova] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Received event network-vif-plugged-a3ce62f1-b1fb-40f1-a551-e9d26d299144 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 832.639140] env[62600]: DEBUG oslo_concurrency.lockutils [req-90bd954c-40c5-461e-b59b-6d1ac311527a req-7c781b2d-7699-48d3-b658-e1cb91864e12 service nova] Acquiring lock "9d899d96-9f4f-41d1-a368-3fde5efc110a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.639380] env[62600]: DEBUG oslo_concurrency.lockutils [req-90bd954c-40c5-461e-b59b-6d1ac311527a req-7c781b2d-7699-48d3-b658-e1cb91864e12 service nova] Lock "9d899d96-9f4f-41d1-a368-3fde5efc110a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.639646] env[62600]: DEBUG oslo_concurrency.lockutils [req-90bd954c-40c5-461e-b59b-6d1ac311527a req-7c781b2d-7699-48d3-b658-e1cb91864e12 service nova] Lock "9d899d96-9f4f-41d1-a368-3fde5efc110a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.639897] env[62600]: DEBUG nova.compute.manager [req-90bd954c-40c5-461e-b59b-6d1ac311527a req-7c781b2d-7699-48d3-b658-e1cb91864e12 service nova] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] No waiting events found dispatching network-vif-plugged-a3ce62f1-b1fb-40f1-a551-e9d26d299144 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 832.640094] env[62600]: WARNING nova.compute.manager [req-90bd954c-40c5-461e-b59b-6d1ac311527a req-7c781b2d-7699-48d3-b658-e1cb91864e12 service nova] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Received unexpected event network-vif-plugged-a3ce62f1-b1fb-40f1-a551-e9d26d299144 for instance with vm_state building and task_state spawning. [ 832.640416] env[62600]: DEBUG nova.compute.manager [req-90bd954c-40c5-461e-b59b-6d1ac311527a req-7c781b2d-7699-48d3-b658-e1cb91864e12 service nova] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Received event network-changed-a3ce62f1-b1fb-40f1-a551-e9d26d299144 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 832.640666] env[62600]: DEBUG nova.compute.manager [req-90bd954c-40c5-461e-b59b-6d1ac311527a req-7c781b2d-7699-48d3-b658-e1cb91864e12 service nova] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Refreshing instance network info cache due to event network-changed-a3ce62f1-b1fb-40f1-a551-e9d26d299144. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 832.640962] env[62600]: DEBUG oslo_concurrency.lockutils [req-90bd954c-40c5-461e-b59b-6d1ac311527a req-7c781b2d-7699-48d3-b658-e1cb91864e12 service nova] Acquiring lock "refresh_cache-9d899d96-9f4f-41d1-a368-3fde5efc110a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.648278] env[62600]: DEBUG oslo_vmware.api [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222464, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.681525] env[62600]: DEBUG nova.scheduler.client.report [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 80 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 832.681525] env[62600]: DEBUG nova.compute.provider_tree [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 80 to 81 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 832.681525] env[62600]: DEBUG nova.compute.provider_tree [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 832.861978] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222462, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.884081] env[62600]: DEBUG nova.network.neutron [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 832.970887] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222463, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.103797] env[62600]: DEBUG nova.network.neutron [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Updating instance_info_cache with network_info: [{"id": "a3ce62f1-b1fb-40f1-a551-e9d26d299144", "address": "fa:16:3e:fd:c2:fa", "network": {"id": "f59f04dc-029f-42be-92de-1d7d3e973fd7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-36149453-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "734eda84fbe4493ba1f3db243a2bdb09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb971244-43ba-41b4-a6a2-a4558548012c", "external-id": "nsx-vlan-transportzone-873", "segmentation_id": 873, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3ce62f1-b1", "ovs_interfaceid": "a3ce62f1-b1fb-40f1-a551-e9d26d299144", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.142518] env[62600]: DEBUG oslo_vmware.api [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222464, 'name': PowerOffVM_Task, 'duration_secs': 0.233903} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.142801] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 833.142973] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 833.143243] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-60827822-46f0-42da-b7e5-20edf01c8c1f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.168400] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 833.168668] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 833.168869] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Deleting the datastore file [datastore2] b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 833.169154] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-28d491c0-128b-4b59-8aee-9655cb3675d0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.175977] env[62600]: DEBUG oslo_vmware.api [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for the task: (returnval){ [ 833.175977] env[62600]: value = "task-1222466" [ 833.175977] env[62600]: _type = "Task" [ 833.175977] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.186056] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.473s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.186232] env[62600]: DEBUG nova.compute.manager [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 833.189026] env[62600]: DEBUG oslo_vmware.api [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222466, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.189400] env[62600]: DEBUG oslo_concurrency.lockutils [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 22.967s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.362752] env[62600]: DEBUG oslo_vmware.api [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222462, 'name': PowerOnVM_Task, 'duration_secs': 0.978645} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.363069] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 833.363277] env[62600]: INFO nova.compute.manager [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Took 6.38 seconds to spawn the instance on the hypervisor. [ 833.363429] env[62600]: DEBUG nova.compute.manager [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 833.364236] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e9843d4-76f5-4259-b854-6ec9614a1ba2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.470046] env[62600]: DEBUG oslo_vmware.api [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222463, 'name': PowerOnVM_Task, 'duration_secs': 0.906722} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.470352] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 833.470571] env[62600]: INFO nova.compute.manager [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Took 8.57 seconds to spawn the instance on the hypervisor. [ 833.470831] env[62600]: DEBUG nova.compute.manager [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 833.471642] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-678ecf8d-bfac-4614-b7c7-90f24df18f9b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.607130] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Releasing lock "refresh_cache-9d899d96-9f4f-41d1-a368-3fde5efc110a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.607501] env[62600]: DEBUG nova.compute.manager [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Instance network_info: |[{"id": "a3ce62f1-b1fb-40f1-a551-e9d26d299144", "address": "fa:16:3e:fd:c2:fa", "network": {"id": "f59f04dc-029f-42be-92de-1d7d3e973fd7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-36149453-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "734eda84fbe4493ba1f3db243a2bdb09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb971244-43ba-41b4-a6a2-a4558548012c", "external-id": "nsx-vlan-transportzone-873", "segmentation_id": 873, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3ce62f1-b1", "ovs_interfaceid": "a3ce62f1-b1fb-40f1-a551-e9d26d299144", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 833.607858] env[62600]: DEBUG oslo_concurrency.lockutils [req-90bd954c-40c5-461e-b59b-6d1ac311527a req-7c781b2d-7699-48d3-b658-e1cb91864e12 service nova] Acquired lock "refresh_cache-9d899d96-9f4f-41d1-a368-3fde5efc110a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.608084] env[62600]: DEBUG nova.network.neutron [req-90bd954c-40c5-461e-b59b-6d1ac311527a req-7c781b2d-7699-48d3-b658-e1cb91864e12 service nova] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Refreshing network info cache for port a3ce62f1-b1fb-40f1-a551-e9d26d299144 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 833.609287] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fd:c2:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cb971244-43ba-41b4-a6a2-a4558548012c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a3ce62f1-b1fb-40f1-a551-e9d26d299144', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 833.616711] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Creating folder: Project (734eda84fbe4493ba1f3db243a2bdb09). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.617243] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-933c4d1c-4402-460f-bccf-d70df5c16bbf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.628400] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Created folder: Project (734eda84fbe4493ba1f3db243a2bdb09) in parent group-v264198. [ 833.628552] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Creating folder: Instances. Parent ref: group-v264277. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.628824] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9c07d7b6-04b8-4719-b5fa-cfdeceb4ca88 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.638013] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Created folder: Instances in parent group-v264277. [ 833.638279] env[62600]: DEBUG oslo.service.loopingcall [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.638479] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 833.638690] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2fb83b48-a280-4c6b-a7cc-0094909affaa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.657399] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 833.657399] env[62600]: value = "task-1222469" [ 833.657399] env[62600]: _type = "Task" [ 833.657399] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.665239] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222469, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.688456] env[62600]: DEBUG oslo_vmware.api [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Task: {'id': task-1222466, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152379} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.689486] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 833.689803] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 833.690114] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 833.690353] env[62600]: INFO nova.compute.manager [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Took 1.07 seconds to destroy the instance on the hypervisor. [ 833.690664] env[62600]: DEBUG oslo.service.loopingcall [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.690920] env[62600]: DEBUG nova.compute.manager [-] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 833.691085] env[62600]: DEBUG nova.network.neutron [-] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 833.697931] env[62600]: DEBUG nova.compute.utils [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 833.701812] env[62600]: DEBUG nova.objects.instance [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lazy-loading 'migration_context' on Instance uuid de4f9304-3357-4eaa-9c94-fe28bc554086 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 833.703123] env[62600]: DEBUG nova.compute.manager [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Not allocating networking since 'none' was specified. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 833.729298] env[62600]: DEBUG nova.network.neutron [-] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 833.731148] env[62600]: DEBUG oslo_concurrency.lockutils [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Acquiring lock "fb6f8e36-8d24-45ea-a6e4-4d768c3b232d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.731148] env[62600]: DEBUG oslo_concurrency.lockutils [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Lock "fb6f8e36-8d24-45ea-a6e4-4d768c3b232d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.731366] env[62600]: DEBUG oslo_concurrency.lockutils [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Acquiring lock "fb6f8e36-8d24-45ea-a6e4-4d768c3b232d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.731506] env[62600]: DEBUG oslo_concurrency.lockutils [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Lock "fb6f8e36-8d24-45ea-a6e4-4d768c3b232d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.731716] env[62600]: DEBUG oslo_concurrency.lockutils [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Lock "fb6f8e36-8d24-45ea-a6e4-4d768c3b232d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.733658] env[62600]: INFO nova.compute.manager [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Terminating instance [ 833.737988] env[62600]: DEBUG nova.compute.manager [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 833.738261] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 833.739210] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a3a8033-27d6-4171-acda-3dac99ff5f85 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.748504] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 833.748504] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e0d6daed-f25b-4d5e-999c-250d33cdd0a6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.755245] env[62600]: DEBUG oslo_vmware.api [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Waiting for the task: (returnval){ [ 833.755245] env[62600]: value = "task-1222470" [ 833.755245] env[62600]: _type = "Task" [ 833.755245] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.763918] env[62600]: DEBUG oslo_vmware.api [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': task-1222470, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.880664] env[62600]: INFO nova.compute.manager [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Took 32.72 seconds to build instance. [ 833.992321] env[62600]: INFO nova.compute.manager [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Took 38.56 seconds to build instance. [ 834.166944] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222469, 'name': CreateVM_Task} progress is 25%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.204913] env[62600]: INFO nova.compute.manager [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Rebuilding instance [ 834.207398] env[62600]: DEBUG nova.compute.manager [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 834.236254] env[62600]: DEBUG nova.network.neutron [-] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.270120] env[62600]: DEBUG oslo_vmware.api [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': task-1222470, 'name': PowerOffVM_Task, 'duration_secs': 0.507735} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.270663] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 834.271532] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 834.271532] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-38d303b0-0274-449d-bdca-42a8a31083e2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.349673] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 834.349932] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 834.350250] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Deleting the datastore file [datastore1] fb6f8e36-8d24-45ea-a6e4-4d768c3b232d {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 834.353255] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f9bf1f1d-7c25-40c0-94e2-a90c4b33c00c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.360900] env[62600]: DEBUG oslo_vmware.api [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Waiting for the task: (returnval){ [ 834.360900] env[62600]: value = "task-1222472" [ 834.360900] env[62600]: _type = "Task" [ 834.360900] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.368905] env[62600]: DEBUG nova.network.neutron [req-90bd954c-40c5-461e-b59b-6d1ac311527a req-7c781b2d-7699-48d3-b658-e1cb91864e12 service nova] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Updated VIF entry in instance network info cache for port a3ce62f1-b1fb-40f1-a551-e9d26d299144. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 834.369278] env[62600]: DEBUG nova.network.neutron [req-90bd954c-40c5-461e-b59b-6d1ac311527a req-7c781b2d-7699-48d3-b658-e1cb91864e12 service nova] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Updating instance_info_cache with network_info: [{"id": "a3ce62f1-b1fb-40f1-a551-e9d26d299144", "address": "fa:16:3e:fd:c2:fa", "network": {"id": "f59f04dc-029f-42be-92de-1d7d3e973fd7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-36149453-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "734eda84fbe4493ba1f3db243a2bdb09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb971244-43ba-41b4-a6a2-a4558548012c", "external-id": "nsx-vlan-transportzone-873", "segmentation_id": 873, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3ce62f1-b1", "ovs_interfaceid": "a3ce62f1-b1fb-40f1-a551-e9d26d299144", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.378309] env[62600]: DEBUG oslo_vmware.api [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': task-1222472, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.382013] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1c2182a3-2df1-4ca2-a408-4dca70118047 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Lock "c2c20e5f-6c0e-4989-bc36-ff4a27d5c534" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.490s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.392970] env[62600]: DEBUG nova.compute.manager [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 834.393873] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81c725a7-63ff-4a21-bbf4-e0533c19996b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.494642] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0b9ce0f5-d812-4b26-947e-a162e93f9345 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "af7036df-b9f0-4ce6-962a-1edd7c1ea211" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.072s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.545023] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cccee40e-d38f-4267-966f-0fe0b618cde6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.553043] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7ac052-8586-475c-98df-92940cdadbf6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.586557] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e79ccbcc-ea9e-4f34-9460-cd352abd9303 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.593934] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2c3f28d-c8c0-4900-9a0e-d495848664df {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.606944] env[62600]: DEBUG nova.compute.provider_tree [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 834.671320] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222469, 'name': CreateVM_Task, 'duration_secs': 0.711087} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.671607] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 834.672664] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.673027] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.673455] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 834.673828] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a216404c-7b1d-4fbd-aa06-86a4f5155b2b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.679780] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 834.679780] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d8ed91-74f4-62aa-f826-c12d63ee69e5" [ 834.679780] env[62600]: _type = "Task" [ 834.679780] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.691186] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d8ed91-74f4-62aa-f826-c12d63ee69e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.738933] env[62600]: INFO nova.compute.manager [-] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Took 1.05 seconds to deallocate network for instance. [ 834.852224] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "af7036df-b9f0-4ce6-962a-1edd7c1ea211" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.852549] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "af7036df-b9f0-4ce6-962a-1edd7c1ea211" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.852789] env[62600]: INFO nova.compute.manager [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Shelving [ 834.870854] env[62600]: DEBUG oslo_vmware.api [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Task: {'id': task-1222472, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133383} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.871209] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 834.871410] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 834.871615] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 834.871848] env[62600]: INFO nova.compute.manager [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 834.872118] env[62600]: DEBUG oslo.service.loopingcall [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.872534] env[62600]: DEBUG oslo_concurrency.lockutils [req-90bd954c-40c5-461e-b59b-6d1ac311527a req-7c781b2d-7699-48d3-b658-e1cb91864e12 service nova] Releasing lock "refresh_cache-9d899d96-9f4f-41d1-a368-3fde5efc110a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.872858] env[62600]: DEBUG nova.compute.manager [-] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 834.872956] env[62600]: DEBUG nova.network.neutron [-] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 834.905352] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 834.905610] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b07fe58d-d8f6-48f0-86f1-607f02c46625 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.912795] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for the task: (returnval){ [ 834.912795] env[62600]: value = "task-1222473" [ 834.912795] env[62600]: _type = "Task" [ 834.912795] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.923847] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222473, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.138463] env[62600]: DEBUG nova.scheduler.client.report [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 81 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 835.138815] env[62600]: DEBUG nova.compute.provider_tree [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 81 to 82 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 835.138926] env[62600]: DEBUG nova.compute.provider_tree [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 835.192032] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d8ed91-74f4-62aa-f826-c12d63ee69e5, 'name': SearchDatastore_Task, 'duration_secs': 0.017399} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.192032] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.192032] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 835.192032] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.192523] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.192523] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 835.192523] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fd5d5810-aa43-4e4a-b1bc-1679f6fe3254 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.201655] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 835.202230] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 835.202698] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-974abc75-3670-49ed-bf71-59a074bf722f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.214555] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 835.214555] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5208f73c-45fe-d10e-56dc-502d1bf02aba" [ 835.214555] env[62600]: _type = "Task" [ 835.214555] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.220118] env[62600]: DEBUG nova.compute.manager [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 835.232148] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5208f73c-45fe-d10e-56dc-502d1bf02aba, 'name': SearchDatastore_Task, 'duration_secs': 0.008589} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.232148] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10593807-9636-4aa6-915e-587d2ea29d83 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.237400] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 835.237400] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5294bacb-8672-35a5-2db3-20bbd858b729" [ 835.237400] env[62600]: _type = "Task" [ 835.237400] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.243735] env[62600]: DEBUG nova.compute.manager [req-a8c2bd11-ffa7-4279-a497-283bea49c55a req-f1a8bacc-dacc-4745-852c-f0c0ae42c9c2 service nova] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Received event network-vif-deleted-6995d562-b51e-4eb4-8d30-6c89c403144d {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.243933] env[62600]: INFO nova.compute.manager [req-a8c2bd11-ffa7-4279-a497-283bea49c55a req-f1a8bacc-dacc-4745-852c-f0c0ae42c9c2 service nova] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Neutron deleted interface 6995d562-b51e-4eb4-8d30-6c89c403144d; detaching it from the instance and deleting it from the info cache [ 835.244109] env[62600]: DEBUG nova.network.neutron [req-a8c2bd11-ffa7-4279-a497-283bea49c55a req-f1a8bacc-dacc-4745-852c-f0c0ae42c9c2 service nova] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.249133] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.249662] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5294bacb-8672-35a5-2db3-20bbd858b729, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.266566] env[62600]: DEBUG nova.virt.hardware [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 835.266829] env[62600]: DEBUG nova.virt.hardware [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 835.267168] env[62600]: DEBUG nova.virt.hardware [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 835.267238] env[62600]: DEBUG nova.virt.hardware [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 835.267335] env[62600]: DEBUG nova.virt.hardware [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 835.267493] env[62600]: DEBUG nova.virt.hardware [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 835.267707] env[62600]: DEBUG nova.virt.hardware [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 835.267875] env[62600]: DEBUG nova.virt.hardware [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 835.268081] env[62600]: DEBUG nova.virt.hardware [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 835.268283] env[62600]: DEBUG nova.virt.hardware [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 835.268467] env[62600]: DEBUG nova.virt.hardware [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 835.269491] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11bbed28-5456-41ee-adad-a95ed163cdab {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.277015] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-261789d7-829b-4d25-bf9c-43f83f8660a6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.290830] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Instance VIF info [] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 835.296452] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Creating folder: Project (4f2885e08aaa4c7f879cd3b8b0e6f8d3). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 835.296740] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c6055213-747a-46c4-8dfb-e642b89a277f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.306836] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Created folder: Project (4f2885e08aaa4c7f879cd3b8b0e6f8d3) in parent group-v264198. [ 835.307115] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Creating folder: Instances. Parent ref: group-v264280. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 835.307348] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ce230dd0-e377-4db1-aad3-ed3f2b11e02e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.316181] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Created folder: Instances in parent group-v264280. [ 835.316414] env[62600]: DEBUG oslo.service.loopingcall [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.316612] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 835.316824] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-97f390be-da7b-47c1-a6aa-4770408e3baa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.334064] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 835.334064] env[62600]: value = "task-1222476" [ 835.334064] env[62600]: _type = "Task" [ 835.334064] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.341162] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222476, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.359856] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 835.360105] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-983b0f45-87e9-430b-8319-a547650e3c13 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.365827] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 835.365827] env[62600]: value = "task-1222477" [ 835.365827] env[62600]: _type = "Task" [ 835.365827] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.373494] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222477, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.424686] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222473, 'name': PowerOffVM_Task, 'duration_secs': 0.143791} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.425371] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 835.425371] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 835.426182] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a108ea73-da2f-400e-9651-91aa1ccafb79 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.433387] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 835.433684] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1b786570-147e-4ac5-986f-b26e9e0615fe {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.457685] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 835.458015] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 835.458267] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Deleting the datastore file [datastore2] c2c20e5f-6c0e-4989-bc36-ff4a27d5c534 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.458617] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-30dae637-6b8b-4c49-9543-9da335713915 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.465608] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for the task: (returnval){ [ 835.465608] env[62600]: value = "task-1222479" [ 835.465608] env[62600]: _type = "Task" [ 835.465608] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.476591] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222479, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.715878] env[62600]: DEBUG nova.network.neutron [-] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.751122] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5294bacb-8672-35a5-2db3-20bbd858b729, 'name': SearchDatastore_Task, 'duration_secs': 0.009579} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.751461] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5eeff2d9-955c-4850-8b13-9a454493b2e6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.753894] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.754264] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 9d899d96-9f4f-41d1-a368-3fde5efc110a/9d899d96-9f4f-41d1-a368-3fde5efc110a.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 835.754593] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ee3b740c-f8c4-460a-a853-40b4c19ee8d2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.764479] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 835.764479] env[62600]: value = "task-1222480" [ 835.764479] env[62600]: _type = "Task" [ 835.764479] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.773043] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf3d8e3-18f5-4ebd-bf04-039001f4b16f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.800095] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222480, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.819547] env[62600]: DEBUG nova.compute.manager [req-a8c2bd11-ffa7-4279-a497-283bea49c55a req-f1a8bacc-dacc-4745-852c-f0c0ae42c9c2 service nova] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Detach interface failed, port_id=6995d562-b51e-4eb4-8d30-6c89c403144d, reason: Instance fb6f8e36-8d24-45ea-a6e4-4d768c3b232d could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 835.849416] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222476, 'name': CreateVM_Task, 'duration_secs': 0.314546} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.849703] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 835.850148] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.850411] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.850759] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 835.851085] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65319e8c-204e-4ab3-a792-33ea9459c8ef {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.857728] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for the task: (returnval){ [ 835.857728] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52adfe3b-46e5-e62d-fe95-3ef74b169be7" [ 835.857728] env[62600]: _type = "Task" [ 835.857728] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.867663] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52adfe3b-46e5-e62d-fe95-3ef74b169be7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.876854] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222477, 'name': PowerOffVM_Task, 'duration_secs': 0.463692} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.877196] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 835.878137] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd965cc-d9f2-45fd-a201-a94677bfb7a0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.899033] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f341944d-a4c3-4e3f-b51e-2df3c188c046 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.980881] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222479, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.125293} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.981199] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 835.981459] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 835.981663] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 836.153832] env[62600]: DEBUG oslo_concurrency.lockutils [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.964s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.161254] env[62600]: DEBUG oslo_concurrency.lockutils [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.731s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.161682] env[62600]: DEBUG nova.objects.instance [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Lazy-loading 'resources' on Instance uuid 58bf5b37-05dd-478f-8c0b-5475ab63bf71 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 836.219885] env[62600]: INFO nova.compute.manager [-] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Took 1.35 seconds to deallocate network for instance. [ 836.278132] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222480, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469414} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.278442] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 9d899d96-9f4f-41d1-a368-3fde5efc110a/9d899d96-9f4f-41d1-a368-3fde5efc110a.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 836.278667] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 836.278942] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18974e68-4f21-4589-838d-e6d0c98c7e55 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.287075] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 836.287075] env[62600]: value = "task-1222481" [ 836.287075] env[62600]: _type = "Task" [ 836.287075] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.296698] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222481, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.368601] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52adfe3b-46e5-e62d-fe95-3ef74b169be7, 'name': SearchDatastore_Task, 'duration_secs': 0.012154} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.368953] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.369221] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 836.369470] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.369623] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.369907] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 836.370203] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-abd8fd86-6bcc-4f16-9112-5938aaab2678 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.378828] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 836.379130] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 836.379778] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1c499f5-1a3d-4132-9627-42132cfe7c30 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.385676] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for the task: (returnval){ [ 836.385676] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52644b62-df0f-ee5f-7cf4-010a5af2ed58" [ 836.385676] env[62600]: _type = "Task" [ 836.385676] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.394286] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52644b62-df0f-ee5f-7cf4-010a5af2ed58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.411634] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Creating Snapshot of the VM instance {{(pid=62600) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 836.411933] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2cc4ca40-a2a6-429d-9045-e22358823d18 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.421050] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 836.421050] env[62600]: value = "task-1222482" [ 836.421050] env[62600]: _type = "Task" [ 836.421050] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.430274] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222482, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.726585] env[62600]: DEBUG oslo_concurrency.lockutils [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.797307] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222481, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064639} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.800390] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 836.801698] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27058015-fa4f-4134-b1cc-d650d1dff73b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.829889] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 9d899d96-9f4f-41d1-a368-3fde5efc110a/9d899d96-9f4f-41d1-a368-3fde5efc110a.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 836.832891] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c1ea6f1-cb19-4d0f-bda1-f5bc37d8ca6b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.859184] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 836.859184] env[62600]: value = "task-1222483" [ 836.859184] env[62600]: _type = "Task" [ 836.859184] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.874107] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222483, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.899623] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52644b62-df0f-ee5f-7cf4-010a5af2ed58, 'name': SearchDatastore_Task, 'duration_secs': 0.009686} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.900517] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95006801-1e81-4a37-a7f3-bfe1e3f69410 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.909076] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for the task: (returnval){ [ 836.909076] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5296b9b0-39cb-9b3e-1b6e-5b1204a6a111" [ 836.909076] env[62600]: _type = "Task" [ 836.909076] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.920015] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5296b9b0-39cb-9b3e-1b6e-5b1204a6a111, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.930112] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222482, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.017936] env[62600]: DEBUG nova.virt.hardware [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 837.018231] env[62600]: DEBUG nova.virt.hardware [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 837.018396] env[62600]: DEBUG nova.virt.hardware [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 837.018584] env[62600]: DEBUG nova.virt.hardware [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 837.018737] env[62600]: DEBUG nova.virt.hardware [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 837.018892] env[62600]: DEBUG nova.virt.hardware [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 837.019126] env[62600]: DEBUG nova.virt.hardware [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 837.019299] env[62600]: DEBUG nova.virt.hardware [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 837.019475] env[62600]: DEBUG nova.virt.hardware [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 837.019647] env[62600]: DEBUG nova.virt.hardware [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 837.019872] env[62600]: DEBUG nova.virt.hardware [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 837.020781] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72853892-0560-4e1c-96fc-cb2ac4b09c6b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.025129] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-866800f6-8f0d-476c-811e-7842638c57eb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.031500] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de61da42-051d-4c1a-acb7-0e695b84d3c1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.038060] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82403ef8-1459-44fd-b311-4ff617feb451 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.049018] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Instance VIF info [] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.054711] env[62600]: DEBUG oslo.service.loopingcall [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.055420] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 837.055662] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0578c165-1817-41b8-8d69-931424636642 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.095145] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2131ada2-7569-45a1-8517-72ad1065cb75 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.100553] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.100553] env[62600]: value = "task-1222484" [ 837.100553] env[62600]: _type = "Task" [ 837.100553] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.106517] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a2eff3-e8f8-4d2f-8ffc-5f41ac3135d8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.116369] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222484, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.124493] env[62600]: DEBUG nova.compute.provider_tree [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.372077] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222483, 'name': ReconfigVM_Task, 'duration_secs': 0.485327} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.372493] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 9d899d96-9f4f-41d1-a368-3fde5efc110a/9d899d96-9f4f-41d1-a368-3fde5efc110a.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 837.373319] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6d6cdd64-15c1-4a5f-b576-950662121f6e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.380568] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 837.380568] env[62600]: value = "task-1222485" [ 837.380568] env[62600]: _type = "Task" [ 837.380568] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.396347] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222485, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.420639] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5296b9b0-39cb-9b3e-1b6e-5b1204a6a111, 'name': SearchDatastore_Task, 'duration_secs': 0.010015} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.420964] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.421277] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] ee7175b9-dbe6-4f90-bd2b-8829194dc6c3/ee7175b9-dbe6-4f90-bd2b-8829194dc6c3.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 837.421563] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-28b86c33-9c89-4e2f-9cc2-52cd14fe3f5d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.432585] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222482, 'name': CreateSnapshot_Task, 'duration_secs': 0.957154} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.434032] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Created Snapshot of the VM instance {{(pid=62600) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 837.434366] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for the task: (returnval){ [ 837.434366] env[62600]: value = "task-1222486" [ 837.434366] env[62600]: _type = "Task" [ 837.434366] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.435206] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0779dfd0-170a-44b9-b2ae-b1d8ec639ee1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.446116] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222486, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.615485] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222484, 'name': CreateVM_Task, 'duration_secs': 0.321324} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.615485] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 837.615911] env[62600]: DEBUG oslo_concurrency.lockutils [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.616167] env[62600]: DEBUG oslo_concurrency.lockutils [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.616562] env[62600]: DEBUG oslo_concurrency.lockutils [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.617329] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3fbe97c-a775-4b85-a31c-854be0a5105c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.624406] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for the task: (returnval){ [ 837.624406] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]524cb7b8-7cb2-f34d-d65e-d47d31cb202c" [ 837.624406] env[62600]: _type = "Task" [ 837.624406] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.628693] env[62600]: DEBUG nova.scheduler.client.report [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 837.638637] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]524cb7b8-7cb2-f34d-d65e-d47d31cb202c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.696690] env[62600]: INFO nova.compute.manager [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Swapping old allocation on dict_keys(['664af347-7147-4bf5-9019-9ae15cb4aa82']) held by migration a871b6f5-809b-4c75-97f7-3f72b2a3dc6a for instance [ 837.729888] env[62600]: DEBUG nova.scheduler.client.report [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Overwriting current allocation {'allocations': {'664af347-7147-4bf5-9019-9ae15cb4aa82': {'resources': {'VCPU': 1, 'MEMORY_MB': 256, 'DISK_GB': 1}, 'generation': 82}}, 'project_id': '2141e7e1a0884ed1b512cf12e0d01e55', 'user_id': '76f3dcf16ac34df683548f101b258060', 'consumer_generation': 1} on consumer de4f9304-3357-4eaa-9c94-fe28bc554086 {{(pid=62600) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 837.847238] env[62600]: DEBUG oslo_concurrency.lockutils [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "refresh_cache-de4f9304-3357-4eaa-9c94-fe28bc554086" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.847528] env[62600]: DEBUG oslo_concurrency.lockutils [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquired lock "refresh_cache-de4f9304-3357-4eaa-9c94-fe28bc554086" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.847800] env[62600]: DEBUG nova.network.neutron [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 837.891239] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222485, 'name': Rename_Task, 'duration_secs': 0.188583} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.891872] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 837.892280] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b8bb4c5f-9936-467a-8734-42870bf3bcda {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.899035] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 837.899035] env[62600]: value = "task-1222487" [ 837.899035] env[62600]: _type = "Task" [ 837.899035] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.908684] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222487, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.948701] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222486, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447318} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.948973] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] ee7175b9-dbe6-4f90-bd2b-8829194dc6c3/ee7175b9-dbe6-4f90-bd2b-8829194dc6c3.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 837.949173] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 837.949441] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-96ea7799-b654-49aa-8377-84c7e454b695 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.959476] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Creating linked-clone VM from snapshot {{(pid=62600) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 837.961025] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-83ab9320-e348-451b-8209-5516b65f6e00 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.964507] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for the task: (returnval){ [ 837.964507] env[62600]: value = "task-1222488" [ 837.964507] env[62600]: _type = "Task" [ 837.964507] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.970599] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 837.970599] env[62600]: value = "task-1222489" [ 837.970599] env[62600]: _type = "Task" [ 837.970599] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.973752] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222488, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.983092] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222489, 'name': CloneVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.136934] env[62600]: DEBUG oslo_concurrency.lockutils [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.976s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.139101] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]524cb7b8-7cb2-f34d-d65e-d47d31cb202c, 'name': SearchDatastore_Task, 'duration_secs': 0.053109} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.140609] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.849s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.142134] env[62600]: INFO nova.compute.claims [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 838.148652] env[62600]: DEBUG oslo_concurrency.lockutils [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.148942] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 838.149246] env[62600]: DEBUG oslo_concurrency.lockutils [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.149419] env[62600]: DEBUG oslo_concurrency.lockutils [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.149541] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 838.150999] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4873ec11-a22d-491d-8d59-245ebd586fa5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.161903] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 838.162812] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 838.164203] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0cbc40f8-497f-4331-a3d7-7b9c71e4c513 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.166310] env[62600]: INFO nova.scheduler.client.report [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Deleted allocations for instance 58bf5b37-05dd-478f-8c0b-5475ab63bf71 [ 838.171189] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for the task: (returnval){ [ 838.171189] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c38394-ba6a-27b0-6f79-68ffe3c6c8dd" [ 838.171189] env[62600]: _type = "Task" [ 838.171189] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.180217] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c38394-ba6a-27b0-6f79-68ffe3c6c8dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.411927] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222487, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.475578] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222488, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068642} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.478947] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.479770] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8a8070-1248-4eb7-a40c-7e071316c61c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.490557] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222489, 'name': CloneVM_Task} progress is 94%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.506269] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] ee7175b9-dbe6-4f90-bd2b-8829194dc6c3/ee7175b9-dbe6-4f90-bd2b-8829194dc6c3.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.506706] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b275ceec-c7ac-49e2-b345-5719c7edbb96 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.529806] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for the task: (returnval){ [ 838.529806] env[62600]: value = "task-1222490" [ 838.529806] env[62600]: _type = "Task" [ 838.529806] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.538777] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222490, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.643109] env[62600]: DEBUG nova.network.neutron [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Updating instance_info_cache with network_info: [{"id": "fe02d314-372f-43fb-8d7c-cfc33c9e74bd", "address": "fa:16:3e:86:22:ec", "network": {"id": "d946b94e-9827-4c22-af05-072fcaa3d0e6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.235", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ec39f2547f624ed0b240d084b949694a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb0e556a-0f69-4a5c-af62-ffc46edb8e63", "external-id": "nsx-vlan-transportzone-136", "segmentation_id": 136, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe02d314-37", "ovs_interfaceid": "fe02d314-372f-43fb-8d7c-cfc33c9e74bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.681957] env[62600]: DEBUG oslo_concurrency.lockutils [None req-44c6caec-c51a-4f9c-9eb0-3e868b6a6166 tempest-InstanceActionsNegativeTestJSON-593075399 tempest-InstanceActionsNegativeTestJSON-593075399-project-member] Lock "58bf5b37-05dd-478f-8c0b-5475ab63bf71" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.189s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.689386] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c38394-ba6a-27b0-6f79-68ffe3c6c8dd, 'name': SearchDatastore_Task, 'duration_secs': 0.010084} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.690410] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17f60fe2-6465-46e7-9162-1161898db56c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.700023] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for the task: (returnval){ [ 838.700023] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ed6aa2-bf0c-5f83-4179-9d418a3b85b2" [ 838.700023] env[62600]: _type = "Task" [ 838.700023] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.712508] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ed6aa2-bf0c-5f83-4179-9d418a3b85b2, 'name': SearchDatastore_Task, 'duration_secs': 0.011078} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.713130] env[62600]: DEBUG oslo_concurrency.lockutils [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.713536] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] c2c20e5f-6c0e-4989-bc36-ff4a27d5c534/c2c20e5f-6c0e-4989-bc36-ff4a27d5c534.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 838.713965] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c60c8c3-85cb-448e-8922-d4f82afbdcfa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.723033] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for the task: (returnval){ [ 838.723033] env[62600]: value = "task-1222491" [ 838.723033] env[62600]: _type = "Task" [ 838.723033] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.732425] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222491, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.911048] env[62600]: DEBUG oslo_vmware.api [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222487, 'name': PowerOnVM_Task, 'duration_secs': 0.980892} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.911365] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 838.911579] env[62600]: INFO nova.compute.manager [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Took 8.30 seconds to spawn the instance on the hypervisor. [ 838.911795] env[62600]: DEBUG nova.compute.manager [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 838.912641] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f430af-1f0b-47c6-b622-16a7c4dda730 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.985824] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222489, 'name': CloneVM_Task} progress is 94%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.040333] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222490, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.146020] env[62600]: DEBUG oslo_concurrency.lockutils [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Releasing lock "refresh_cache-de4f9304-3357-4eaa-9c94-fe28bc554086" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.146693] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 839.147109] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-22d826e3-8c5a-4d7e-b16a-85e8b1f0b138 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.160132] env[62600]: DEBUG oslo_vmware.api [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 839.160132] env[62600]: value = "task-1222492" [ 839.160132] env[62600]: _type = "Task" [ 839.160132] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.176814] env[62600]: DEBUG oslo_vmware.api [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222492, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.243444] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222491, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.433163] env[62600]: INFO nova.compute.manager [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Took 33.38 seconds to build instance. [ 839.486599] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222489, 'name': CloneVM_Task} progress is 95%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.534017] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2422344-bca9-49ca-b3c6-78649030c5f2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.546711] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222490, 'name': ReconfigVM_Task, 'duration_secs': 0.58699} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.549125] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Reconfigured VM instance instance-00000040 to attach disk [datastore1] ee7175b9-dbe6-4f90-bd2b-8829194dc6c3/ee7175b9-dbe6-4f90-bd2b-8829194dc6c3.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.551115] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a2c72e22-6cbb-4017-a117-66b0d2deed94 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.552909] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc17cc3-6c7a-4990-9e10-aa6d6698dc17 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.591828] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4906e2cf-4f00-49cb-a60c-a2013aaa7cc6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.595649] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for the task: (returnval){ [ 839.595649] env[62600]: value = "task-1222493" [ 839.595649] env[62600]: _type = "Task" [ 839.595649] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.605014] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f324a240-94ed-428c-a62b-2e3203464c3b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.612910] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222493, 'name': Rename_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.626634] env[62600]: DEBUG nova.compute.provider_tree [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.678019] env[62600]: DEBUG oslo_vmware.api [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222492, 'name': PowerOffVM_Task, 'duration_secs': 0.295644} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.678486] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 839.679202] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T12:01:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='99810421-2db4-4272-82bd-17ba5175b496',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-2077855957',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 839.679432] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 839.679604] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 839.679816] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 839.679992] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 839.680171] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 839.680385] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 839.680560] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 839.680758] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 839.680974] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 839.681222] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 839.686429] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c31ad764-a28e-4c25-99f9-6be2efdb1258 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.703692] env[62600]: DEBUG oslo_vmware.api [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 839.703692] env[62600]: value = "task-1222494" [ 839.703692] env[62600]: _type = "Task" [ 839.703692] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.713010] env[62600]: DEBUG oslo_vmware.api [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222494, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.734304] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222491, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.689983} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.734579] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] c2c20e5f-6c0e-4989-bc36-ff4a27d5c534/c2c20e5f-6c0e-4989-bc36-ff4a27d5c534.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 839.734798] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 839.735067] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fd710168-58f9-41cc-a31b-37a381566ab4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.742899] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for the task: (returnval){ [ 839.742899] env[62600]: value = "task-1222495" [ 839.742899] env[62600]: _type = "Task" [ 839.742899] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.754110] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222495, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.935802] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb7648ef-18ce-4629-96db-f3adcee11577 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "9d899d96-9f4f-41d1-a368-3fde5efc110a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.635s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.990756] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222489, 'name': CloneVM_Task, 'duration_secs': 1.720477} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.991805] env[62600]: INFO nova.virt.vmwareapi.vmops [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Created linked-clone VM from snapshot [ 839.992702] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d2d110-bd69-453e-b4b8-46539d1b0aa7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.003104] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Uploading image 0a198f1c-06cc-49b4-8be7-28919c48706d {{(pid=62600) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 840.027871] env[62600]: DEBUG oslo_vmware.rw_handles [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 840.027871] env[62600]: value = "vm-264285" [ 840.027871] env[62600]: _type = "VirtualMachine" [ 840.027871] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 840.028199] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-049dea87-fc5d-4da2-a89d-c6b4426a6819 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.036276] env[62600]: DEBUG oslo_vmware.rw_handles [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lease: (returnval){ [ 840.036276] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5206b425-f064-8e78-bf4e-3cc57b2d265d" [ 840.036276] env[62600]: _type = "HttpNfcLease" [ 840.036276] env[62600]: } obtained for exporting VM: (result){ [ 840.036276] env[62600]: value = "vm-264285" [ 840.036276] env[62600]: _type = "VirtualMachine" [ 840.036276] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 840.036527] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the lease: (returnval){ [ 840.036527] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5206b425-f064-8e78-bf4e-3cc57b2d265d" [ 840.036527] env[62600]: _type = "HttpNfcLease" [ 840.036527] env[62600]: } to be ready. {{(pid=62600) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 840.043856] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 840.043856] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5206b425-f064-8e78-bf4e-3cc57b2d265d" [ 840.043856] env[62600]: _type = "HttpNfcLease" [ 840.043856] env[62600]: } is initializing. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 840.115036] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222493, 'name': Rename_Task, 'duration_secs': 0.250195} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.116404] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 840.116692] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a21e7766-519f-4c5e-8e62-4c8f3ba49156 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.125887] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for the task: (returnval){ [ 840.125887] env[62600]: value = "task-1222497" [ 840.125887] env[62600]: _type = "Task" [ 840.125887] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.133711] env[62600]: DEBUG nova.scheduler.client.report [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 840.142706] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222497, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.214393] env[62600]: DEBUG oslo_vmware.api [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222494, 'name': ReconfigVM_Task, 'duration_secs': 0.268496} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.215356] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-832b7663-8a28-4599-862d-5f864308f1ee {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.240084] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T12:01:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='99810421-2db4-4272-82bd-17ba5175b496',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-2077855957',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 840.240255] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 840.240425] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 840.240614] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 840.240787] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 840.240936] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 840.241312] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 840.241951] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 840.241951] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 840.241951] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 840.242232] env[62600]: DEBUG nova.virt.hardware [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 840.242904] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32eddd7e-b3ae-4de3-bade-cbeede1eaf7d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.253751] env[62600]: DEBUG oslo_vmware.api [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 840.253751] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d3e80d-e399-4433-80ad-57786633546e" [ 840.253751] env[62600]: _type = "Task" [ 840.253751] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.255892] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222495, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069125} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.258928] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 840.259830] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa54403e-08e6-40c7-851f-f0cdb6075dcc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.281896] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] c2c20e5f-6c0e-4989-bc36-ff4a27d5c534/c2c20e5f-6c0e-4989-bc36-ff4a27d5c534.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 840.283933] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cabd373f-165d-4c41-8975-1b2ea893eede {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.299546] env[62600]: DEBUG oslo_vmware.api [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d3e80d-e399-4433-80ad-57786633546e, 'name': SearchDatastore_Task, 'duration_secs': 0.013245} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.305008] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Reconfiguring VM instance instance-0000002b to detach disk 2000 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 840.307503] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbe5977e-defd-4dca-a11f-09b3ee42ec71 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.324926] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for the task: (returnval){ [ 840.324926] env[62600]: value = "task-1222498" [ 840.324926] env[62600]: _type = "Task" [ 840.324926] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.332503] env[62600]: DEBUG oslo_vmware.api [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 840.332503] env[62600]: value = "task-1222499" [ 840.332503] env[62600]: _type = "Task" [ 840.332503] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.337720] env[62600]: DEBUG oslo_vmware.rw_handles [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5244e429-4da1-4429-476a-de6d515f9670/disk-0.vmdk. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 840.338202] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222498, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.339925] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fbbbb94-7669-4e81-973b-d0d3375764e0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.350287] env[62600]: DEBUG oslo_vmware.rw_handles [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5244e429-4da1-4429-476a-de6d515f9670/disk-0.vmdk is in state: ready. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 840.350512] env[62600]: ERROR oslo_vmware.rw_handles [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5244e429-4da1-4429-476a-de6d515f9670/disk-0.vmdk due to incomplete transfer. [ 840.354089] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-68d57d34-9d82-404a-85b4-a32420d0c1d3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.355933] env[62600]: DEBUG oslo_vmware.api [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222499, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.364249] env[62600]: DEBUG oslo_vmware.rw_handles [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5244e429-4da1-4429-476a-de6d515f9670/disk-0.vmdk. {{(pid=62600) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 840.364249] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Uploaded image 2685df45-5a39-4d23-9105-ef28747ec7d4 to the Glance image server {{(pid=62600) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 840.366093] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Destroying the VM {{(pid=62600) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 840.366466] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-99d4e5ba-9736-4c37-8927-aebe2d5bf7c9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.375020] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 840.375020] env[62600]: value = "task-1222500" [ 840.375020] env[62600]: _type = "Task" [ 840.375020] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.386193] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222500, 'name': Destroy_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.547050] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 840.547050] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5206b425-f064-8e78-bf4e-3cc57b2d265d" [ 840.547050] env[62600]: _type = "HttpNfcLease" [ 840.547050] env[62600]: } is ready. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 840.547050] env[62600]: DEBUG oslo_vmware.rw_handles [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 840.547050] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5206b425-f064-8e78-bf4e-3cc57b2d265d" [ 840.547050] env[62600]: _type = "HttpNfcLease" [ 840.547050] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 840.547050] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-306031b8-7631-4c58-b2e5-8ad2b41be229 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.556418] env[62600]: DEBUG oslo_vmware.rw_handles [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5259e3c9-3002-cfef-14f0-fd71bdeb7cc8/disk-0.vmdk from lease info. {{(pid=62600) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 840.557160] env[62600]: DEBUG oslo_vmware.rw_handles [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5259e3c9-3002-cfef-14f0-fd71bdeb7cc8/disk-0.vmdk for reading. {{(pid=62600) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 840.641023] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.641023] env[62600]: DEBUG nova.compute.manager [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 840.644627] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.230s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.644926] env[62600]: DEBUG nova.objects.instance [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Lazy-loading 'resources' on Instance uuid 734e8a2b-faaa-4165-b0e9-69eeb3699b2d {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 840.653608] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222497, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.706045] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b849cad0-6acd-4c6e-98d9-fdd4712f52bd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.835841] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222498, 'name': ReconfigVM_Task, 'duration_secs': 0.311033} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.836072] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Reconfigured VM instance instance-0000003e to attach disk [datastore1] c2c20e5f-6c0e-4989-bc36-ff4a27d5c534/c2c20e5f-6c0e-4989-bc36-ff4a27d5c534.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 840.836729] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-215c05f3-b35b-46a0-a711-92b85d7ff970 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.847057] env[62600]: DEBUG oslo_vmware.api [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222499, 'name': ReconfigVM_Task, 'duration_secs': 0.321646} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.848456] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Reconfigured VM instance instance-0000002b to detach disk 2000 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 840.848712] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for the task: (returnval){ [ 840.848712] env[62600]: value = "task-1222501" [ 840.848712] env[62600]: _type = "Task" [ 840.848712] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.849468] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5ac274f-df31-4579-bc39-073608cbdbb1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.860833] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222501, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.878431] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] de4f9304-3357-4eaa-9c94-fe28bc554086/de4f9304-3357-4eaa-9c94-fe28bc554086.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 840.878857] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f89c2a57-19c0-4057-b90e-2703c5c73034 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.901688] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222500, 'name': Destroy_Task} progress is 33%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.903142] env[62600]: DEBUG oslo_vmware.api [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 840.903142] env[62600]: value = "task-1222502" [ 840.903142] env[62600]: _type = "Task" [ 840.903142] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.911872] env[62600]: DEBUG oslo_vmware.api [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222502, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.985445] env[62600]: DEBUG nova.compute.manager [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 840.986570] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b4a64a8-fceb-4b87-a049-4806e025ec47 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.137426] env[62600]: DEBUG oslo_vmware.api [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222497, 'name': PowerOnVM_Task, 'duration_secs': 0.735184} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.137955] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 841.138307] env[62600]: INFO nova.compute.manager [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Took 5.92 seconds to spawn the instance on the hypervisor. [ 841.138625] env[62600]: DEBUG nova.compute.manager [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.139515] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44ef08a8-a080-46b5-b19a-abfe1095a919 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.153023] env[62600]: DEBUG nova.compute.utils [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 841.153023] env[62600]: DEBUG nova.compute.manager [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 841.153023] env[62600]: DEBUG nova.network.neutron [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 841.212383] env[62600]: DEBUG nova.policy [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a2416189869c4cbe950d6fe3d4310eb8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '48e1772c9a6c4671ab89df2e8b40cae1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 841.363569] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222501, 'name': Rename_Task, 'duration_secs': 0.160824} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.364079] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 841.364538] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d9f68b43-2977-4fc9-a971-ec201e4e4697 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.377248] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for the task: (returnval){ [ 841.377248] env[62600]: value = "task-1222503" [ 841.377248] env[62600]: _type = "Task" [ 841.377248] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.392407] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222503, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.398167] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222500, 'name': Destroy_Task, 'duration_secs': 0.591848} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.399582] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Destroyed the VM [ 841.399582] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Deleting Snapshot of the VM instance {{(pid=62600) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 841.399582] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-7e11d64a-64f6-4a65-91d7-9116451d2975 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.411743] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 841.411743] env[62600]: value = "task-1222504" [ 841.411743] env[62600]: _type = "Task" [ 841.411743] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.420935] env[62600]: DEBUG oslo_vmware.api [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222502, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.432292] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222504, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.498530] env[62600]: INFO nova.compute.manager [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] instance snapshotting [ 841.501608] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da33a1e-cbb4-4d81-8ecc-7ddf0809881f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.525314] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e244dda-71a1-421a-b191-7c86122071fa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.552894] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1513406-717c-4399-9d54-553ed9473933 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.562839] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6c37ffc-1329-456e-b03e-1a617cbbaba2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.600289] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a7a9af-9b69-4313-9f81-f52dc3f97b8d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.610676] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0b20a49-d40a-4e30-a943-3a6ed218a7e2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.627844] env[62600]: DEBUG nova.compute.provider_tree [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 841.651084] env[62600]: DEBUG nova.network.neutron [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Successfully created port: d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 841.661487] env[62600]: DEBUG nova.compute.manager [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 841.667626] env[62600]: INFO nova.compute.manager [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Took 33.44 seconds to build instance. [ 841.887860] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222503, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.915930] env[62600]: DEBUG oslo_vmware.api [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222502, 'name': ReconfigVM_Task, 'duration_secs': 0.561014} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.922319] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Reconfigured VM instance instance-0000002b to attach disk [datastore2] de4f9304-3357-4eaa-9c94-fe28bc554086/de4f9304-3357-4eaa-9c94-fe28bc554086.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 841.923435] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1de37309-bd7c-400f-80fc-9031100c4800 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.933312] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222504, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.951632] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bfff086-891b-4163-b646-6f43c075442a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.976634] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d212d2b6-82b0-496c-88b1-eec39b1b5230 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.999262] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e71d5744-bd84-4bc9-94f3-3a314ab458aa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.008152] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.008451] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-423ef3dc-3038-4313-bd35-5a2d98b99ba4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.017322] env[62600]: DEBUG oslo_vmware.api [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 842.017322] env[62600]: value = "task-1222505" [ 842.017322] env[62600]: _type = "Task" [ 842.017322] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.026581] env[62600]: DEBUG oslo_vmware.api [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222505, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.038346] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Creating Snapshot of the VM instance {{(pid=62600) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 842.038703] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e8259eec-3edc-42b7-8539-0b1d5382dd32 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.049347] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 842.049347] env[62600]: value = "task-1222506" [ 842.049347] env[62600]: _type = "Task" [ 842.049347] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.059242] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222506, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.154827] env[62600]: ERROR nova.scheduler.client.report [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] [req-dcf1ff7b-aed6-4d02-9fca-aa7d57d06645] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 664af347-7147-4bf5-9019-9ae15cb4aa82. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-dcf1ff7b-aed6-4d02-9fca-aa7d57d06645"}]} [ 842.172825] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5c6533e1-4ca3-4a26-874c-b1e0aa24574a tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Lock "ee7175b9-dbe6-4f90-bd2b-8829194dc6c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.046s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.175856] env[62600]: DEBUG nova.scheduler.client.report [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Refreshing inventories for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 842.197751] env[62600]: DEBUG nova.scheduler.client.report [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Updating ProviderTree inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 842.198009] env[62600]: DEBUG nova.compute.provider_tree [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 842.205317] env[62600]: INFO nova.compute.manager [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Rebuilding instance [ 842.215956] env[62600]: DEBUG nova.scheduler.client.report [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Refreshing aggregate associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, aggregates: None {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 842.242229] env[62600]: DEBUG nova.scheduler.client.report [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Refreshing trait associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 842.266130] env[62600]: DEBUG nova.compute.manager [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 842.267221] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5fad0ac-3a25-4249-977d-990ebf000a44 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.389564] env[62600]: DEBUG oslo_vmware.api [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222503, 'name': PowerOnVM_Task, 'duration_secs': 0.841054} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.389673] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 842.389898] env[62600]: DEBUG nova.compute.manager [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 842.390715] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90926f7-116b-40be-a791-988a403a0783 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.438821] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222504, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.531093] env[62600]: DEBUG oslo_vmware.api [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222505, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.560325] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222506, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.586842] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01382161-6eab-4225-8325-ff0554241751 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.595811] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b0c336-9192-40de-b1f1-59011b1c56a6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.631088] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d2a96c-a5de-4ea7-81c0-eb1bb0b9dbf1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.641942] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8534ca2-a4c5-4ef0-b658-b44f913b0778 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.660584] env[62600]: DEBUG nova.compute.provider_tree [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 842.671269] env[62600]: DEBUG nova.compute.manager [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 842.698520] env[62600]: DEBUG nova.virt.hardware [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 842.698816] env[62600]: DEBUG nova.virt.hardware [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 842.699072] env[62600]: DEBUG nova.virt.hardware [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 842.699546] env[62600]: DEBUG nova.virt.hardware [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 842.699679] env[62600]: DEBUG nova.virt.hardware [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 842.699950] env[62600]: DEBUG nova.virt.hardware [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 842.700227] env[62600]: DEBUG nova.virt.hardware [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 842.700421] env[62600]: DEBUG nova.virt.hardware [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 842.700613] env[62600]: DEBUG nova.virt.hardware [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 842.700808] env[62600]: DEBUG nova.virt.hardware [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 842.701018] env[62600]: DEBUG nova.virt.hardware [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 842.701971] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64111480-f8fe-4308-8110-2a0a9d24a1da {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.712273] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef79488e-68ef-4374-87e0-acf33a6cd62a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.782486] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 842.783204] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-70bdf782-89ae-4726-bae2-b5728292851f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.792737] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for the task: (returnval){ [ 842.792737] env[62600]: value = "task-1222507" [ 842.792737] env[62600]: _type = "Task" [ 842.792737] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.803827] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222507, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.910751] env[62600]: DEBUG oslo_concurrency.lockutils [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.925922] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222504, 'name': RemoveSnapshot_Task} progress is 50%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.031251] env[62600]: DEBUG oslo_vmware.api [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222505, 'name': PowerOnVM_Task, 'duration_secs': 0.810097} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.031919] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 843.061116] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222506, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.196645] env[62600]: DEBUG nova.scheduler.client.report [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 83 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 843.196645] env[62600]: DEBUG nova.compute.provider_tree [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 83 to 84 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 843.196645] env[62600]: DEBUG nova.compute.provider_tree [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 843.262961] env[62600]: DEBUG nova.compute.manager [req-9cbbe28e-5d60-49d1-8f51-222115055954 req-831f1325-5b2c-4a86-9f12-835f8bf5eb60 service nova] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Received event network-vif-plugged-d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 843.262961] env[62600]: DEBUG oslo_concurrency.lockutils [req-9cbbe28e-5d60-49d1-8f51-222115055954 req-831f1325-5b2c-4a86-9f12-835f8bf5eb60 service nova] Acquiring lock "b967fb11-e70e-4e17-b769-38da581bd83b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.262961] env[62600]: DEBUG oslo_concurrency.lockutils [req-9cbbe28e-5d60-49d1-8f51-222115055954 req-831f1325-5b2c-4a86-9f12-835f8bf5eb60 service nova] Lock "b967fb11-e70e-4e17-b769-38da581bd83b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.262961] env[62600]: DEBUG oslo_concurrency.lockutils [req-9cbbe28e-5d60-49d1-8f51-222115055954 req-831f1325-5b2c-4a86-9f12-835f8bf5eb60 service nova] Lock "b967fb11-e70e-4e17-b769-38da581bd83b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.262961] env[62600]: DEBUG nova.compute.manager [req-9cbbe28e-5d60-49d1-8f51-222115055954 req-831f1325-5b2c-4a86-9f12-835f8bf5eb60 service nova] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] No waiting events found dispatching network-vif-plugged-d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 843.263212] env[62600]: WARNING nova.compute.manager [req-9cbbe28e-5d60-49d1-8f51-222115055954 req-831f1325-5b2c-4a86-9f12-835f8bf5eb60 service nova] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Received unexpected event network-vif-plugged-d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049 for instance with vm_state building and task_state spawning. [ 843.308025] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222507, 'name': PowerOffVM_Task, 'duration_secs': 0.166341} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.308025] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 843.308025] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 843.308025] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e68706-ffca-4583-a9be-dd505fafc86b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.317472] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 843.317472] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d0fff5a9-493b-4d05-9758-f1bc07479bcf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.348227] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 843.348227] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 843.348227] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Deleting the datastore file [datastore1] ee7175b9-dbe6-4f90-bd2b-8829194dc6c3 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 843.348563] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d12f1116-f587-4517-9141-6c4ddabcae49 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.360169] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for the task: (returnval){ [ 843.360169] env[62600]: value = "task-1222509" [ 843.360169] env[62600]: _type = "Task" [ 843.360169] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.370224] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222509, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.381388] env[62600]: DEBUG nova.network.neutron [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Successfully updated port: d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 843.428418] env[62600]: DEBUG oslo_vmware.api [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222504, 'name': RemoveSnapshot_Task, 'duration_secs': 1.737346} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.429120] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Deleted Snapshot of the VM instance {{(pid=62600) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 843.429379] env[62600]: INFO nova.compute.manager [None req-cc12a415-ce86-4bca-a149-2d38ad8b8f07 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Took 19.64 seconds to snapshot the instance on the hypervisor. [ 843.561821] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222506, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.642049] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Acquiring lock "c2c20e5f-6c0e-4989-bc36-ff4a27d5c534" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.642049] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Lock "c2c20e5f-6c0e-4989-bc36-ff4a27d5c534" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.642049] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Acquiring lock "c2c20e5f-6c0e-4989-bc36-ff4a27d5c534-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.642049] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Lock "c2c20e5f-6c0e-4989-bc36-ff4a27d5c534-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.642334] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Lock "c2c20e5f-6c0e-4989-bc36-ff4a27d5c534-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.645179] env[62600]: INFO nova.compute.manager [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Terminating instance [ 843.647459] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Acquiring lock "refresh_cache-c2c20e5f-6c0e-4989-bc36-ff4a27d5c534" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.647736] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Acquired lock "refresh_cache-c2c20e5f-6c0e-4989-bc36-ff4a27d5c534" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.648017] env[62600]: DEBUG nova.network.neutron [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 843.701631] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.057s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.704430] env[62600]: DEBUG oslo_concurrency.lockutils [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.934s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.704671] env[62600]: DEBUG nova.objects.instance [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Lazy-loading 'resources' on Instance uuid 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 843.722651] env[62600]: INFO nova.scheduler.client.report [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Deleted allocations for instance 734e8a2b-faaa-4165-b0e9-69eeb3699b2d [ 843.868817] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222509, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.320572} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.869115] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 843.869312] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 843.869497] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 843.884140] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "refresh_cache-b967fb11-e70e-4e17-b769-38da581bd83b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.884291] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquired lock "refresh_cache-b967fb11-e70e-4e17-b769-38da581bd83b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.884747] env[62600]: DEBUG nova.network.neutron [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 844.041885] env[62600]: INFO nova.compute.manager [None req-007b79e2-7324-4b59-ad22-f4fe9a0dad0e tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Updating instance to original state: 'active' [ 844.064178] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222506, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.172093] env[62600]: DEBUG nova.network.neutron [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 844.232132] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c394129b-3ad4-4785-ba32-fdadaf03a7b8 tempest-AttachInterfacesV270Test-1184226025 tempest-AttachInterfacesV270Test-1184226025-project-member] Lock "734e8a2b-faaa-4165-b0e9-69eeb3699b2d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.759s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.263548] env[62600]: DEBUG nova.network.neutron [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.437241] env[62600]: DEBUG nova.network.neutron [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 844.540044] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fd20305-3e6c-44df-9441-d05e407f57e8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.562297] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954f7080-8fc5-4196-8edf-d0d7906d58e2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.580980] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222506, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.609089] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e9c71a-95ad-4bb3-b67b-c0728c5b90af {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.619328] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2478508-8b45-4a98-8e4f-f2c4f9766b4a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.639816] env[62600]: DEBUG nova.compute.provider_tree [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.688605] env[62600]: DEBUG nova.network.neutron [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Updating instance_info_cache with network_info: [{"id": "d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049", "address": "fa:16:3e:4f:fd:77", "network": {"id": "036dc114-6ea0-46cb-a958-4a6137e7e85f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-728627542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e1772c9a6c4671ab89df2e8b40cae1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6e6a2fb-7b", "ovs_interfaceid": "d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.770222] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Releasing lock "refresh_cache-c2c20e5f-6c0e-4989-bc36-ff4a27d5c534" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.770798] env[62600]: DEBUG nova.compute.manager [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 844.771065] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 844.772899] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567402ee-ee07-48b6-8ae9-89efacbd0ce1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.782583] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 844.782928] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-69aba452-c0f7-4f45-9b56-80eaf82dadbd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.793077] env[62600]: DEBUG oslo_vmware.api [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for the task: (returnval){ [ 844.793077] env[62600]: value = "task-1222510" [ 844.793077] env[62600]: _type = "Task" [ 844.793077] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.803823] env[62600]: DEBUG oslo_vmware.api [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222510, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.907958] env[62600]: DEBUG nova.virt.hardware [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 844.908270] env[62600]: DEBUG nova.virt.hardware [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 844.908509] env[62600]: DEBUG nova.virt.hardware [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 844.908820] env[62600]: DEBUG nova.virt.hardware [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 844.908980] env[62600]: DEBUG nova.virt.hardware [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 844.909222] env[62600]: DEBUG nova.virt.hardware [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 844.909502] env[62600]: DEBUG nova.virt.hardware [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 844.909705] env[62600]: DEBUG nova.virt.hardware [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 844.909981] env[62600]: DEBUG nova.virt.hardware [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 844.910244] env[62600]: DEBUG nova.virt.hardware [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 844.910518] env[62600]: DEBUG nova.virt.hardware [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 844.912349] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a96c71c2-ac08-4ec0-8d7c-15b368f21691 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.922745] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96fc16b0-6dc5-456d-92c8-17fca2f51e9a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.940376] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Instance VIF info [] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.946205] env[62600]: DEBUG oslo.service.loopingcall [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.946516] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 844.946750] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e7333f5-d1fc-4763-a169-fa5afaa3cd79 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.966292] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.966292] env[62600]: value = "task-1222511" [ 844.966292] env[62600]: _type = "Task" [ 844.966292] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.975749] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222511, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.088973] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222506, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.146268] env[62600]: DEBUG nova.scheduler.client.report [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.192256] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Releasing lock "refresh_cache-b967fb11-e70e-4e17-b769-38da581bd83b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.192625] env[62600]: DEBUG nova.compute.manager [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Instance network_info: |[{"id": "d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049", "address": "fa:16:3e:4f:fd:77", "network": {"id": "036dc114-6ea0-46cb-a958-4a6137e7e85f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-728627542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e1772c9a6c4671ab89df2e8b40cae1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6e6a2fb-7b", "ovs_interfaceid": "d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 845.193195] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:fd:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3e05affa-2640-435e-a124-0ee8a6ab1152', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd6e6a2fb-7b6b-42bd-8b1a-c8026b45d049', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 845.205018] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Creating folder: Project (48e1772c9a6c4671ab89df2e8b40cae1). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 845.205018] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-63b3bc32-4a23-4387-9514-cfee2fad5466 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.216704] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Created folder: Project (48e1772c9a6c4671ab89df2e8b40cae1) in parent group-v264198. [ 845.216921] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Creating folder: Instances. Parent ref: group-v264288. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 845.217477] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-45922f9d-9a48-4a19-acf0-844483c24186 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.229348] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Created folder: Instances in parent group-v264288. [ 845.232219] env[62600]: DEBUG oslo.service.loopingcall [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.232219] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 845.232219] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-94d22c9c-7f15-4903-971c-52abf2329ed2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.252960] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 845.252960] env[62600]: value = "task-1222514" [ 845.252960] env[62600]: _type = "Task" [ 845.252960] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.263642] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222514, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.304864] env[62600]: DEBUG oslo_vmware.api [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222510, 'name': PowerOffVM_Task, 'duration_secs': 0.128207} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.305194] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 845.305384] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 845.305648] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-27908a03-1a1c-449b-9a4f-4f7aa58f56c5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.339174] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 845.339321] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 845.339591] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Deleting the datastore file [datastore1] c2c20e5f-6c0e-4989-bc36-ff4a27d5c534 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 845.339948] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4ca72e3e-7fb2-47cc-a838-992237b7c83b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.348488] env[62600]: DEBUG oslo_vmware.api [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for the task: (returnval){ [ 845.348488] env[62600]: value = "task-1222516" [ 845.348488] env[62600]: _type = "Task" [ 845.348488] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.357845] env[62600]: DEBUG oslo_vmware.api [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222516, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.448400] env[62600]: DEBUG nova.compute.manager [req-888f0e43-6a50-4708-a4d8-d56ab800965d req-4e9e1996-e069-40ab-8364-605bb33fb414 service nova] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Received event network-changed-d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 845.448400] env[62600]: DEBUG nova.compute.manager [req-888f0e43-6a50-4708-a4d8-d56ab800965d req-4e9e1996-e069-40ab-8364-605bb33fb414 service nova] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Refreshing instance network info cache due to event network-changed-d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 845.449055] env[62600]: DEBUG oslo_concurrency.lockutils [req-888f0e43-6a50-4708-a4d8-d56ab800965d req-4e9e1996-e069-40ab-8364-605bb33fb414 service nova] Acquiring lock "refresh_cache-b967fb11-e70e-4e17-b769-38da581bd83b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.449308] env[62600]: DEBUG oslo_concurrency.lockutils [req-888f0e43-6a50-4708-a4d8-d56ab800965d req-4e9e1996-e069-40ab-8364-605bb33fb414 service nova] Acquired lock "refresh_cache-b967fb11-e70e-4e17-b769-38da581bd83b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.449408] env[62600]: DEBUG nova.network.neutron [req-888f0e43-6a50-4708-a4d8-d56ab800965d req-4e9e1996-e069-40ab-8364-605bb33fb414 service nova] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Refreshing network info cache for port d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 845.480373] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222511, 'name': CreateVM_Task, 'duration_secs': 0.401396} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.480685] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 845.481246] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.481434] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.481770] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 845.482055] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00b17578-420b-44bf-b0cb-997fdd2ffc1c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.492591] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for the task: (returnval){ [ 845.492591] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d6c716-bca3-dcbd-d43c-04365817a02f" [ 845.492591] env[62600]: _type = "Task" [ 845.492591] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.503273] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d6c716-bca3-dcbd-d43c-04365817a02f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.584031] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222506, 'name': CreateSnapshot_Task, 'duration_secs': 3.150944} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.584328] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Created Snapshot of the VM instance {{(pid=62600) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 845.585650] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf63e56-dceb-436a-8013-347c17297915 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.653343] env[62600]: DEBUG oslo_concurrency.lockutils [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.949s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.657238] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.193s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.659812] env[62600]: INFO nova.compute.claims [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.679983] env[62600]: INFO nova.scheduler.client.report [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Deleted allocations for instance 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d [ 845.755335] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "de4f9304-3357-4eaa-9c94-fe28bc554086" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.755335] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "de4f9304-3357-4eaa-9c94-fe28bc554086" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.755335] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "de4f9304-3357-4eaa-9c94-fe28bc554086-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.755508] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "de4f9304-3357-4eaa-9c94-fe28bc554086-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.755707] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "de4f9304-3357-4eaa-9c94-fe28bc554086-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.761830] env[62600]: INFO nova.compute.manager [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Terminating instance [ 845.764060] env[62600]: DEBUG nova.compute.manager [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 845.764279] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 845.765121] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e715955-c2b7-462d-b661-89a43ada3265 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.771906] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222514, 'name': CreateVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.777655] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 845.777969] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-427c558e-94a2-4ba0-af35-da6c754d2016 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.786231] env[62600]: DEBUG oslo_vmware.api [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 845.786231] env[62600]: value = "task-1222517" [ 845.786231] env[62600]: _type = "Task" [ 845.786231] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.796436] env[62600]: DEBUG oslo_vmware.api [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222517, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.858739] env[62600]: DEBUG oslo_vmware.api [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Task: {'id': task-1222516, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.107568} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.859417] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 845.859417] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 845.859417] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 845.859593] env[62600]: INFO nova.compute.manager [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Took 1.09 seconds to destroy the instance on the hypervisor. [ 845.859820] env[62600]: DEBUG oslo.service.loopingcall [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.860109] env[62600]: DEBUG nova.compute.manager [-] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 845.860148] env[62600]: DEBUG nova.network.neutron [-] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 845.876729] env[62600]: DEBUG nova.network.neutron [-] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 846.007351] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d6c716-bca3-dcbd-d43c-04365817a02f, 'name': SearchDatastore_Task, 'duration_secs': 0.01087} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.007655] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.007907] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 846.008175] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.008330] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.008546] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 846.008786] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1f141c1f-dfdb-4069-9042-61e9ff2a7d41 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.020662] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 846.021502] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 846.022781] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75d37537-d70c-405b-811c-b79459527c7d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.032949] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for the task: (returnval){ [ 846.032949] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b2a646-05f0-8419-40b0-9cec6d9183e3" [ 846.032949] env[62600]: _type = "Task" [ 846.032949] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.042699] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b2a646-05f0-8419-40b0-9cec6d9183e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.105963] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Creating linked-clone VM from snapshot {{(pid=62600) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 846.109468] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-884bc99b-7e96-4846-bb4e-50adddd67a15 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.120149] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 846.120149] env[62600]: value = "task-1222518" [ 846.120149] env[62600]: _type = "Task" [ 846.120149] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.129035] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222518, 'name': CloneVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.188792] env[62600]: DEBUG oslo_concurrency.lockutils [None req-13ffd49b-d218-4c59-bb0c-e8bf1706b77a tempest-SecurityGroupsTestJSON-1599770198 tempest-SecurityGroupsTestJSON-1599770198-project-member] Lock "92bd8b10-2034-49d4-bf38-0edc7e1e6f7d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.308s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.272613] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222514, 'name': CreateVM_Task, 'duration_secs': 0.558111} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.275320] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 846.277443] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.277443] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.277443] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 846.279503] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63b53147-b1cf-4ecf-aaae-a8fb9be282e7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.285228] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 846.285228] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521080f2-476f-24c4-0818-01ac61cc42c4" [ 846.285228] env[62600]: _type = "Task" [ 846.285228] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.303098] env[62600]: DEBUG oslo_vmware.api [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222517, 'name': PowerOffVM_Task, 'duration_secs': 0.466788} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.304108] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521080f2-476f-24c4-0818-01ac61cc42c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.304108] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 846.304108] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 846.304412] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-60fd3b27-e911-4b7f-981f-88f3526cd532 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.380152] env[62600]: DEBUG nova.network.neutron [-] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.382524] env[62600]: DEBUG nova.network.neutron [req-888f0e43-6a50-4708-a4d8-d56ab800965d req-4e9e1996-e069-40ab-8364-605bb33fb414 service nova] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Updated VIF entry in instance network info cache for port d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 846.382749] env[62600]: DEBUG nova.network.neutron [req-888f0e43-6a50-4708-a4d8-d56ab800965d req-4e9e1996-e069-40ab-8364-605bb33fb414 service nova] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Updating instance_info_cache with network_info: [{"id": "d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049", "address": "fa:16:3e:4f:fd:77", "network": {"id": "036dc114-6ea0-46cb-a958-4a6137e7e85f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-728627542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e1772c9a6c4671ab89df2e8b40cae1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6e6a2fb-7b", "ovs_interfaceid": "d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.385174] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 846.385446] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 846.385613] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Deleting the datastore file [datastore2] de4f9304-3357-4eaa-9c94-fe28bc554086 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 846.386295] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-98e58d82-5e77-4079-be8b-b4e30037042d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.396994] env[62600]: DEBUG oslo_vmware.api [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for the task: (returnval){ [ 846.396994] env[62600]: value = "task-1222520" [ 846.396994] env[62600]: _type = "Task" [ 846.396994] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.408020] env[62600]: DEBUG oslo_vmware.api [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222520, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.529587] env[62600]: DEBUG oslo_concurrency.lockutils [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquiring lock "a358a3cb-deda-419a-aa3c-ce7aeb534240" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.530041] env[62600]: DEBUG oslo_concurrency.lockutils [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lock "a358a3cb-deda-419a-aa3c-ce7aeb534240" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.530313] env[62600]: DEBUG oslo_concurrency.lockutils [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquiring lock "a358a3cb-deda-419a-aa3c-ce7aeb534240-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.530599] env[62600]: DEBUG oslo_concurrency.lockutils [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lock "a358a3cb-deda-419a-aa3c-ce7aeb534240-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.530909] env[62600]: DEBUG oslo_concurrency.lockutils [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lock "a358a3cb-deda-419a-aa3c-ce7aeb534240-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.534231] env[62600]: INFO nova.compute.manager [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Terminating instance [ 846.540303] env[62600]: DEBUG nova.compute.manager [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 846.540649] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 846.541657] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ad3b391-7523-4a83-9085-2853313b897c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.552744] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b2a646-05f0-8419-40b0-9cec6d9183e3, 'name': SearchDatastore_Task, 'duration_secs': 0.016897} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.555851] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 846.556107] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-803da385-a341-4023-b616-bfd5f487d0fb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.561765] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-020ba76a-9376-4944-b948-e63d2e72fed2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.566709] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for the task: (returnval){ [ 846.566709] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]529aa81d-f7c4-26ea-51e8-73e599441f8b" [ 846.566709] env[62600]: _type = "Task" [ 846.566709] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.575154] env[62600]: DEBUG oslo_vmware.api [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 846.575154] env[62600]: value = "task-1222521" [ 846.575154] env[62600]: _type = "Task" [ 846.575154] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.582843] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]529aa81d-f7c4-26ea-51e8-73e599441f8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.588866] env[62600]: DEBUG oslo_vmware.api [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222521, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.632475] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222518, 'name': CloneVM_Task} progress is 94%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.807053] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521080f2-476f-24c4-0818-01ac61cc42c4, 'name': SearchDatastore_Task, 'duration_secs': 0.0197} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.810329] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.810625] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 846.810870] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.889302] env[62600]: INFO nova.compute.manager [-] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Took 1.03 seconds to deallocate network for instance. [ 846.889777] env[62600]: DEBUG oslo_concurrency.lockutils [req-888f0e43-6a50-4708-a4d8-d56ab800965d req-4e9e1996-e069-40ab-8364-605bb33fb414 service nova] Releasing lock "refresh_cache-b967fb11-e70e-4e17-b769-38da581bd83b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.908720] env[62600]: DEBUG oslo_vmware.api [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Task: {'id': task-1222520, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172448} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.911968] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 846.911968] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 846.911968] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 846.911968] env[62600]: INFO nova.compute.manager [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Took 1.15 seconds to destroy the instance on the hypervisor. [ 846.912298] env[62600]: DEBUG oslo.service.loopingcall [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.912678] env[62600]: DEBUG nova.compute.manager [-] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 846.912775] env[62600]: DEBUG nova.network.neutron [-] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 847.021140] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-671d8418-4c3c-47f8-804b-50f1c1264040 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.035163] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f68481-b39c-4b24-9a89-d78c7c715b9e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.072079] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-328f7edd-a7e9-4e36-a239-e62d7b53a28d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.084874] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]529aa81d-f7c4-26ea-51e8-73e599441f8b, 'name': SearchDatastore_Task, 'duration_secs': 0.031003} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.085629] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.085907] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] ee7175b9-dbe6-4f90-bd2b-8829194dc6c3/ee7175b9-dbe6-4f90-bd2b-8829194dc6c3.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 847.087166] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1ecdf36-91b1-421a-b8a2-15fab2fef981 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.096103] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.096364] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 847.096848] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5f89e552-0fdc-4fac-8ba7-f1031067ae38 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.098778] env[62600]: DEBUG oslo_vmware.api [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222521, 'name': PowerOffVM_Task, 'duration_secs': 0.203998} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.098947] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-577aac16-9ca6-4eeb-bbfe-60d3acde6701 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.103197] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 847.103384] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 847.103979] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4459881c-6ded-4cbf-a9a2-028bbeab1b3a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.115584] env[62600]: DEBUG nova.compute.provider_tree [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 847.120308] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for the task: (returnval){ [ 847.120308] env[62600]: value = "task-1222522" [ 847.120308] env[62600]: _type = "Task" [ 847.120308] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.120587] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 847.120788] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 847.121903] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b33e8516-77aa-4676-abac-cae919c25d6c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.138320] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 847.138320] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5204edc9-844e-03ee-a14c-01544c707799" [ 847.138320] env[62600]: _type = "Task" [ 847.138320] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.144675] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222518, 'name': CloneVM_Task} progress is 94%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.145354] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222522, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.155909] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5204edc9-844e-03ee-a14c-01544c707799, 'name': SearchDatastore_Task, 'duration_secs': 0.012777} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.156872] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa34c45c-4092-4c2a-89f7-ff1b793e9915 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.164809] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 847.164809] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52697c43-8791-b9f8-094d-b75328240c35" [ 847.164809] env[62600]: _type = "Task" [ 847.164809] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.175372] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52697c43-8791-b9f8-094d-b75328240c35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.194349] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 847.194607] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 847.194977] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Deleting the datastore file [datastore1] a358a3cb-deda-419a-aa3c-ce7aeb534240 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 847.195347] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-41dd5b44-7600-408c-b34b-5744307d8dd0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.203440] env[62600]: DEBUG oslo_vmware.api [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 847.203440] env[62600]: value = "task-1222524" [ 847.203440] env[62600]: _type = "Task" [ 847.203440] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.213101] env[62600]: DEBUG oslo_vmware.api [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222524, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.397701] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.639854] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222522, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.653426] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222518, 'name': CloneVM_Task} progress is 95%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.662973] env[62600]: DEBUG nova.scheduler.client.report [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 84 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 847.663394] env[62600]: DEBUG nova.compute.provider_tree [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 84 to 85 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 847.663705] env[62600]: DEBUG nova.compute.provider_tree [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 847.680717] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52697c43-8791-b9f8-094d-b75328240c35, 'name': SearchDatastore_Task, 'duration_secs': 0.011939} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.682434] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.682742] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] b967fb11-e70e-4e17-b769-38da581bd83b/b967fb11-e70e-4e17-b769-38da581bd83b.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 847.684408] env[62600]: DEBUG nova.compute.manager [req-027426ae-c57b-4946-999c-25545c26e34b req-0915df12-1ef0-42cd-a2b6-d6ba6988723e service nova] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Received event network-vif-deleted-fe02d314-372f-43fb-8d7c-cfc33c9e74bd {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.684694] env[62600]: INFO nova.compute.manager [req-027426ae-c57b-4946-999c-25545c26e34b req-0915df12-1ef0-42cd-a2b6-d6ba6988723e service nova] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Neutron deleted interface fe02d314-372f-43fb-8d7c-cfc33c9e74bd; detaching it from the instance and deleting it from the info cache [ 847.684783] env[62600]: DEBUG nova.network.neutron [req-027426ae-c57b-4946-999c-25545c26e34b req-0915df12-1ef0-42cd-a2b6-d6ba6988723e service nova] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.686436] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-275711d3-462c-4804-bfe2-a17c84066292 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.696613] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 847.696613] env[62600]: value = "task-1222525" [ 847.696613] env[62600]: _type = "Task" [ 847.696613] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.709484] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222525, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.720348] env[62600]: DEBUG oslo_vmware.api [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222524, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.208878} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.720552] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 847.720740] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 847.720927] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 847.721196] env[62600]: INFO nova.compute.manager [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Took 1.18 seconds to destroy the instance on the hypervisor. [ 847.721477] env[62600]: DEBUG oslo.service.loopingcall [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 847.721676] env[62600]: DEBUG nova.compute.manager [-] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 847.721774] env[62600]: DEBUG nova.network.neutron [-] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 847.911031] env[62600]: DEBUG nova.network.neutron [-] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.144566] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222518, 'name': CloneVM_Task, 'duration_secs': 1.709087} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.149950] env[62600]: INFO nova.virt.vmwareapi.vmops [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Created linked-clone VM from snapshot [ 848.150334] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222522, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.151454] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94790417-ebaf-43dc-9714-e37659771060 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.163288] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Uploading image e1395438-00bb-4e48-add8-ef3068f3c1c3 {{(pid=62600) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 848.169559] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.513s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.170302] env[62600]: DEBUG nova.compute.manager [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 848.173399] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.239s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.173998] env[62600]: DEBUG nova.objects.instance [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Lazy-loading 'resources' on Instance uuid 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 848.187059] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-24748301-7298-45f3-9c18-316642a6f666 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.199100] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24efc4b7-e4de-4372-a1df-5a70d69055b2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.213049] env[62600]: DEBUG oslo_vmware.rw_handles [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 848.213049] env[62600]: value = "vm-264291" [ 848.213049] env[62600]: _type = "VirtualMachine" [ 848.213049] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 848.213431] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-876342eb-4187-4442-b48a-a3caed57aee3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.224772] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222525, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.229259] env[62600]: DEBUG oslo_vmware.rw_handles [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lease: (returnval){ [ 848.229259] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5273912a-11e0-0a60-727a-7b6a82acdec1" [ 848.229259] env[62600]: _type = "HttpNfcLease" [ 848.229259] env[62600]: } obtained for exporting VM: (result){ [ 848.229259] env[62600]: value = "vm-264291" [ 848.229259] env[62600]: _type = "VirtualMachine" [ 848.229259] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 848.229259] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the lease: (returnval){ [ 848.229259] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5273912a-11e0-0a60-727a-7b6a82acdec1" [ 848.229259] env[62600]: _type = "HttpNfcLease" [ 848.229259] env[62600]: } to be ready. {{(pid=62600) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 848.240718] env[62600]: DEBUG nova.compute.manager [req-027426ae-c57b-4946-999c-25545c26e34b req-0915df12-1ef0-42cd-a2b6-d6ba6988723e service nova] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Detach interface failed, port_id=fe02d314-372f-43fb-8d7c-cfc33c9e74bd, reason: Instance de4f9304-3357-4eaa-9c94-fe28bc554086 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 848.245886] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 848.245886] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5273912a-11e0-0a60-727a-7b6a82acdec1" [ 848.245886] env[62600]: _type = "HttpNfcLease" [ 848.245886] env[62600]: } is ready. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 848.245886] env[62600]: DEBUG oslo_vmware.rw_handles [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 848.245886] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5273912a-11e0-0a60-727a-7b6a82acdec1" [ 848.245886] env[62600]: _type = "HttpNfcLease" [ 848.245886] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 848.246225] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65aaab23-3138-44f9-803d-15612a45ccac {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.254838] env[62600]: DEBUG oslo_vmware.rw_handles [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524febcb-0961-8ec1-6050-b89178972c4b/disk-0.vmdk from lease info. {{(pid=62600) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 848.254921] env[62600]: DEBUG oslo_vmware.rw_handles [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524febcb-0961-8ec1-6050-b89178972c4b/disk-0.vmdk for reading. {{(pid=62600) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 848.416167] env[62600]: INFO nova.compute.manager [-] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Took 1.50 seconds to deallocate network for instance. [ 848.426157] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0e41de3b-3e23-4209-8738-e3ced71b97f9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.643403] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222522, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.656050] env[62600]: DEBUG nova.network.neutron [-] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.677231] env[62600]: DEBUG nova.compute.utils [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 848.681799] env[62600]: DEBUG nova.compute.manager [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 848.681799] env[62600]: DEBUG nova.network.neutron [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 848.726307] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222525, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.788517] env[62600]: DEBUG nova.policy [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a2416189869c4cbe950d6fe3d4310eb8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '48e1772c9a6c4671ab89df2e8b40cae1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 848.852575] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquiring lock "afe838cc-e086-4986-87ec-4e1266bcaf60" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.852575] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "afe838cc-e086-4986-87ec-4e1266bcaf60" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.924511] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.067324] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa9d0766-50ec-4987-bea6-093cb6ebb588 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.078171] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-294604ef-27c2-493f-abcf-f6a61664641c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.120413] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9332d2ff-26ec-41bc-9fb0-e4a12419685f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.132362] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dee14e8-dd23-4ba6-a4fb-9cc7c4dd24db {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.144886] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222522, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.536733} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.145767] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] ee7175b9-dbe6-4f90-bd2b-8829194dc6c3/ee7175b9-dbe6-4f90-bd2b-8829194dc6c3.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 849.146100] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 849.146420] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f6ef139c-ceb4-4f00-9fdf-dcc24d266a6b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.165958] env[62600]: INFO nova.compute.manager [-] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Took 1.44 seconds to deallocate network for instance. [ 849.166802] env[62600]: DEBUG nova.compute.provider_tree [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 849.176020] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for the task: (returnval){ [ 849.176020] env[62600]: value = "task-1222527" [ 849.176020] env[62600]: _type = "Task" [ 849.176020] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.186640] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222527, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.188172] env[62600]: DEBUG nova.compute.manager [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 849.229314] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222525, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.404649} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.229637] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] b967fb11-e70e-4e17-b769-38da581bd83b/b967fb11-e70e-4e17-b769-38da581bd83b.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 849.229865] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 849.231133] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-42fec1d5-3798-4539-9a36-66d6d2cc4bec {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.244620] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 849.244620] env[62600]: value = "task-1222528" [ 849.244620] env[62600]: _type = "Task" [ 849.244620] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.258884] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222528, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.355640] env[62600]: DEBUG oslo_vmware.rw_handles [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5259e3c9-3002-cfef-14f0-fd71bdeb7cc8/disk-0.vmdk. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 849.357071] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b27d2cca-ac82-4436-85a7-9315034eb6f7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.361447] env[62600]: DEBUG nova.compute.utils [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 849.367723] env[62600]: DEBUG oslo_vmware.rw_handles [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5259e3c9-3002-cfef-14f0-fd71bdeb7cc8/disk-0.vmdk is in state: ready. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 849.368807] env[62600]: ERROR oslo_vmware.rw_handles [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5259e3c9-3002-cfef-14f0-fd71bdeb7cc8/disk-0.vmdk due to incomplete transfer. [ 849.368807] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-962ba9e9-19e7-45a4-9f95-b18a480982fb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.377426] env[62600]: DEBUG oslo_vmware.rw_handles [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5259e3c9-3002-cfef-14f0-fd71bdeb7cc8/disk-0.vmdk. {{(pid=62600) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 849.377765] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Uploaded image 0a198f1c-06cc-49b4-8be7-28919c48706d to the Glance image server {{(pid=62600) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 849.380402] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Destroying the VM {{(pid=62600) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 849.380928] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-3b1e8c34-5569-4afb-8011-2fc3e036425c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.391457] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 849.391457] env[62600]: value = "task-1222529" [ 849.391457] env[62600]: _type = "Task" [ 849.391457] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.401965] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222529, 'name': Destroy_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.661136] env[62600]: DEBUG nova.network.neutron [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Successfully created port: 57606b61-13bc-4562-b28d-d61f215809e0 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 849.674017] env[62600]: DEBUG nova.scheduler.client.report [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 849.682681] env[62600]: DEBUG oslo_concurrency.lockutils [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.689929] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222527, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07337} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.691043] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 849.691829] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcc0e264-69d8-459f-85a9-60cbcedd3540 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.721999] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] ee7175b9-dbe6-4f90-bd2b-8829194dc6c3/ee7175b9-dbe6-4f90-bd2b-8829194dc6c3.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 849.723523] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1fde7b41-ee66-4f98-a80d-b0c45b1efd43 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.749230] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for the task: (returnval){ [ 849.749230] env[62600]: value = "task-1222530" [ 849.749230] env[62600]: _type = "Task" [ 849.749230] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.767354] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222530, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.767615] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222528, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07402} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.768132] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 849.769111] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d251a69d-4ab8-446b-b64a-c86d3892c42a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.794344] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] b967fb11-e70e-4e17-b769-38da581bd83b/b967fb11-e70e-4e17-b769-38da581bd83b.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 849.794941] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-22d40082-7c03-40f6-bc65-83319529640e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.822032] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 849.822032] env[62600]: value = "task-1222531" [ 849.822032] env[62600]: _type = "Task" [ 849.822032] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.832665] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222531, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.865774] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "afe838cc-e086-4986-87ec-4e1266bcaf60" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.013s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.902572] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222529, 'name': Destroy_Task} progress is 33%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.177967] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.004s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.180665] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.134s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.181160] env[62600]: DEBUG nova.objects.instance [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Lazy-loading 'resources' on Instance uuid 92cf1acb-b471-4add-a73f-7e9e94fbaaa4 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 850.202814] env[62600]: DEBUG nova.compute.manager [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 850.213038] env[62600]: INFO nova.scheduler.client.report [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Deleted allocations for instance 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd [ 850.239175] env[62600]: DEBUG nova.virt.hardware [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 850.239567] env[62600]: DEBUG nova.virt.hardware [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 850.240191] env[62600]: DEBUG nova.virt.hardware [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 850.240988] env[62600]: DEBUG nova.virt.hardware [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 850.240988] env[62600]: DEBUG nova.virt.hardware [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 850.241162] env[62600]: DEBUG nova.virt.hardware [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 850.242043] env[62600]: DEBUG nova.virt.hardware [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 850.242043] env[62600]: DEBUG nova.virt.hardware [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 850.242043] env[62600]: DEBUG nova.virt.hardware [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 850.242043] env[62600]: DEBUG nova.virt.hardware [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 850.242043] env[62600]: DEBUG nova.virt.hardware [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 850.242840] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47e0b9e7-fcd1-4320-a4a2-6b07c3c9caba {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.258735] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b94df181-9dc8-4238-9c1b-49c2ad8680c0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.270125] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222530, 'name': ReconfigVM_Task, 'duration_secs': 0.468396} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.279918] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Reconfigured VM instance instance-00000040 to attach disk [datastore2] ee7175b9-dbe6-4f90-bd2b-8829194dc6c3/ee7175b9-dbe6-4f90-bd2b-8829194dc6c3.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.281759] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5d844f13-fa32-411f-aa8e-a9a4dc98f377 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.289069] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for the task: (returnval){ [ 850.289069] env[62600]: value = "task-1222532" [ 850.289069] env[62600]: _type = "Task" [ 850.289069] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.299975] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222532, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.331508] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222531, 'name': ReconfigVM_Task, 'duration_secs': 0.464628} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.331947] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Reconfigured VM instance instance-00000041 to attach disk [datastore2] b967fb11-e70e-4e17-b769-38da581bd83b/b967fb11-e70e-4e17-b769-38da581bd83b.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.332824] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6367312b-6eb3-478c-ac71-38c0a0f86592 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.343825] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 850.343825] env[62600]: value = "task-1222533" [ 850.343825] env[62600]: _type = "Task" [ 850.343825] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.358145] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222533, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.404404] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222529, 'name': Destroy_Task, 'duration_secs': 0.622931} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.405073] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Destroyed the VM [ 850.405462] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Deleting Snapshot of the VM instance {{(pid=62600) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 850.405855] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-32bf62a6-557d-4774-9c69-e20211bd8c51 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.417008] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 850.417008] env[62600]: value = "task-1222534" [ 850.417008] env[62600]: _type = "Task" [ 850.417008] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.426062] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222534, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.682752] env[62600]: DEBUG nova.compute.manager [req-c6c1a4a2-6f4e-4a37-8ffc-d88f4f0283a8 req-94fe48f4-c5e2-4941-9805-19aa3115b074 service nova] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Received event network-vif-deleted-6a3323e4-5bec-4c2d-8c89-285faea1dd06 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.727974] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8f0145f6-ab6b-4a4e-9172-0928417b9c9c tempest-ServerMetadataNegativeTestJSON-728139754 tempest-ServerMetadataNegativeTestJSON-728139754-project-member] Lock "8f3e87a5-bb70-4e50-9fea-6bf53946b8dd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.856s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.802993] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222532, 'name': Rename_Task, 'duration_secs': 0.187136} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.804783] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 850.804783] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f3086620-3558-4f49-96eb-dc8e4a786cdb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.812361] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for the task: (returnval){ [ 850.812361] env[62600]: value = "task-1222535" [ 850.812361] env[62600]: _type = "Task" [ 850.812361] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.825563] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222535, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.863543] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222533, 'name': Rename_Task, 'duration_secs': 0.192624} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.863863] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 850.864188] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fb4b7a53-257a-4d9b-8e59-5978a40779ed {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.872887] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 850.872887] env[62600]: value = "task-1222536" [ 850.872887] env[62600]: _type = "Task" [ 850.872887] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.888250] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222536, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.927944] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222534, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.944304] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquiring lock "afe838cc-e086-4986-87ec-4e1266bcaf60" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.944623] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "afe838cc-e086-4986-87ec-4e1266bcaf60" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.944858] env[62600]: INFO nova.compute.manager [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Attaching volume 11e471b3-6f58-476c-a94e-4572bdbb8269 to /dev/sdb [ 850.998425] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce863824-c629-4f56-a63b-52485866a7de {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.012040] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e121de49-41ee-46ed-91d4-a59c4c5e689d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.036612] env[62600]: DEBUG nova.virt.block_device [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Updating existing volume attachment record: f0b0778e-ca2a-477c-9bff-9f2a3d7f5f2f {{(pid=62600) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 851.084783] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c625f8-df31-4cc0-987c-99876f37ea4d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.094050] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e126b5f8-cdf1-455a-bda8-23f858115638 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.134662] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c2feb92-d63c-4909-9e3d-d96f1f4a1c25 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.144703] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1fa47a1-b009-4f00-b406-a707bbc4814c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.161629] env[62600]: DEBUG nova.compute.provider_tree [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.325947] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222535, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.384039] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222536, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.434075] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222534, 'name': RemoveSnapshot_Task, 'duration_secs': 0.565556} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.434075] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Deleted Snapshot of the VM instance {{(pid=62600) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 851.434075] env[62600]: DEBUG nova.compute.manager [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.434825] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ff402e-226f-4c0e-8e8d-d835b9190c7b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.665141] env[62600]: DEBUG nova.scheduler.client.report [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 851.825725] env[62600]: DEBUG oslo_vmware.api [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222535, 'name': PowerOnVM_Task, 'duration_secs': 0.646373} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.826233] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 851.826292] env[62600]: DEBUG nova.compute.manager [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.827438] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e9d8ba-f853-42fa-8051-23b9e87a565b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.843555] env[62600]: DEBUG nova.network.neutron [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Successfully updated port: 57606b61-13bc-4562-b28d-d61f215809e0 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 851.864696] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "1b856aff-66be-4fa5-b8b7-124ac24fb4d1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.864696] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "1b856aff-66be-4fa5-b8b7-124ac24fb4d1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.885561] env[62600]: DEBUG oslo_vmware.api [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222536, 'name': PowerOnVM_Task, 'duration_secs': 0.662082} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.885839] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 851.886056] env[62600]: INFO nova.compute.manager [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Took 9.21 seconds to spawn the instance on the hypervisor. [ 851.886240] env[62600]: DEBUG nova.compute.manager [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.887171] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-818ff48f-caf3-4d5b-8654-cf6e8e695394 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.952804] env[62600]: INFO nova.compute.manager [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Shelve offloading [ 851.954993] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 851.955278] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e00b3a0-f8a0-47e2-8353-590d7ce18774 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.965172] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 851.965172] env[62600]: value = "task-1222540" [ 851.965172] env[62600]: _type = "Task" [ 851.965172] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.975699] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] VM already powered off {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 851.975918] env[62600]: DEBUG nova.compute.manager [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.976794] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a86c84a-772d-4dd1-b87d-b7c5070bb13b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.984505] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "refresh_cache-af7036df-b9f0-4ce6-962a-1edd7c1ea211" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.984726] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired lock "refresh_cache-af7036df-b9f0-4ce6-962a-1edd7c1ea211" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.985180] env[62600]: DEBUG nova.network.neutron [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 852.172971] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.992s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.175659] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.921s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.177482] env[62600]: INFO nova.compute.claims [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 852.205250] env[62600]: INFO nova.scheduler.client.report [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Deleted allocations for instance 92cf1acb-b471-4add-a73f-7e9e94fbaaa4 [ 852.349944] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "refresh_cache-d66c404d-8fd1-4fb7-a3b9-f21854c7e735" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.349944] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquired lock "refresh_cache-d66c404d-8fd1-4fb7-a3b9-f21854c7e735" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.349944] env[62600]: DEBUG nova.network.neutron [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 852.357404] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.369328] env[62600]: DEBUG nova.compute.manager [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 852.416099] env[62600]: INFO nova.compute.manager [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Took 41.15 seconds to build instance. [ 852.716214] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d2299f0-a3a8-40eb-ae11-986938da1920 tempest-ServersTestBootFromVolume-190206476 tempest-ServersTestBootFromVolume-190206476-project-member] Lock "92cf1acb-b471-4add-a73f-7e9e94fbaaa4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.013s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.886733] env[62600]: DEBUG nova.compute.manager [req-63aa84ca-5792-43ee-9bfe-f52d6e77598a req-976a2721-9340-45fc-9bf2-0e60c6210577 service nova] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Received event network-vif-plugged-57606b61-13bc-4562-b28d-d61f215809e0 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.886959] env[62600]: DEBUG oslo_concurrency.lockutils [req-63aa84ca-5792-43ee-9bfe-f52d6e77598a req-976a2721-9340-45fc-9bf2-0e60c6210577 service nova] Acquiring lock "d66c404d-8fd1-4fb7-a3b9-f21854c7e735-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.887305] env[62600]: DEBUG oslo_concurrency.lockutils [req-63aa84ca-5792-43ee-9bfe-f52d6e77598a req-976a2721-9340-45fc-9bf2-0e60c6210577 service nova] Lock "d66c404d-8fd1-4fb7-a3b9-f21854c7e735-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.887522] env[62600]: DEBUG oslo_concurrency.lockutils [req-63aa84ca-5792-43ee-9bfe-f52d6e77598a req-976a2721-9340-45fc-9bf2-0e60c6210577 service nova] Lock "d66c404d-8fd1-4fb7-a3b9-f21854c7e735-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.887802] env[62600]: DEBUG nova.compute.manager [req-63aa84ca-5792-43ee-9bfe-f52d6e77598a req-976a2721-9340-45fc-9bf2-0e60c6210577 service nova] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] No waiting events found dispatching network-vif-plugged-57606b61-13bc-4562-b28d-d61f215809e0 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 852.887932] env[62600]: WARNING nova.compute.manager [req-63aa84ca-5792-43ee-9bfe-f52d6e77598a req-976a2721-9340-45fc-9bf2-0e60c6210577 service nova] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Received unexpected event network-vif-plugged-57606b61-13bc-4562-b28d-d61f215809e0 for instance with vm_state building and task_state spawning. [ 852.888199] env[62600]: DEBUG nova.compute.manager [req-63aa84ca-5792-43ee-9bfe-f52d6e77598a req-976a2721-9340-45fc-9bf2-0e60c6210577 service nova] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Received event network-changed-57606b61-13bc-4562-b28d-d61f215809e0 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.888397] env[62600]: DEBUG nova.compute.manager [req-63aa84ca-5792-43ee-9bfe-f52d6e77598a req-976a2721-9340-45fc-9bf2-0e60c6210577 service nova] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Refreshing instance network info cache due to event network-changed-57606b61-13bc-4562-b28d-d61f215809e0. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 852.888621] env[62600]: DEBUG oslo_concurrency.lockutils [req-63aa84ca-5792-43ee-9bfe-f52d6e77598a req-976a2721-9340-45fc-9bf2-0e60c6210577 service nova] Acquiring lock "refresh_cache-d66c404d-8fd1-4fb7-a3b9-f21854c7e735" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.903745] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.906362] env[62600]: DEBUG nova.network.neutron [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 852.919099] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d4e30d69-5b34-4db9-bfc0-7dcbab9fabf0 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "b967fb11-e70e-4e17-b769-38da581bd83b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.837s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.176207] env[62600]: DEBUG nova.network.neutron [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Updating instance_info_cache with network_info: [{"id": "57606b61-13bc-4562-b28d-d61f215809e0", "address": "fa:16:3e:53:05:09", "network": {"id": "036dc114-6ea0-46cb-a958-4a6137e7e85f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-728627542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e1772c9a6c4671ab89df2e8b40cae1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57606b61-13", "ovs_interfaceid": "57606b61-13bc-4562-b28d-d61f215809e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.277146] env[62600]: DEBUG nova.network.neutron [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Updating instance_info_cache with network_info: [{"id": "b15ca3aa-0a63-4385-9171-98748d4b988c", "address": "fa:16:3e:0d:e9:b6", "network": {"id": "5ef67e8a-5613-4cdb-9f35-0a18720657ff", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1644754500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bab74e3ae78248909dbcd483abbb8da2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb15ca3aa-0a", "ovs_interfaceid": "b15ca3aa-0a63-4385-9171-98748d4b988c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.485654] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c5c6d4c-590e-4320-bec8-1aa64e2b4917 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.497311] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78eb4b7a-2279-4233-b506-a00e83385b2c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.540202] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad51645-f849-4b9c-8d3f-a523a5f6a47c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.549608] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe25899-89f0-4438-b640-1d9124e677bb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.565419] env[62600]: DEBUG nova.compute.provider_tree [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.657405] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Acquiring lock "ee7175b9-dbe6-4f90-bd2b-8829194dc6c3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.657723] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Lock "ee7175b9-dbe6-4f90-bd2b-8829194dc6c3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.658016] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Acquiring lock "ee7175b9-dbe6-4f90-bd2b-8829194dc6c3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.658310] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Lock "ee7175b9-dbe6-4f90-bd2b-8829194dc6c3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.658518] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Lock "ee7175b9-dbe6-4f90-bd2b-8829194dc6c3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.661524] env[62600]: INFO nova.compute.manager [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Terminating instance [ 853.663756] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Acquiring lock "refresh_cache-ee7175b9-dbe6-4f90-bd2b-8829194dc6c3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.663935] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Acquired lock "refresh_cache-ee7175b9-dbe6-4f90-bd2b-8829194dc6c3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.664145] env[62600]: DEBUG nova.network.neutron [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 853.678489] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Releasing lock "refresh_cache-d66c404d-8fd1-4fb7-a3b9-f21854c7e735" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.678842] env[62600]: DEBUG nova.compute.manager [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Instance network_info: |[{"id": "57606b61-13bc-4562-b28d-d61f215809e0", "address": "fa:16:3e:53:05:09", "network": {"id": "036dc114-6ea0-46cb-a958-4a6137e7e85f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-728627542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e1772c9a6c4671ab89df2e8b40cae1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57606b61-13", "ovs_interfaceid": "57606b61-13bc-4562-b28d-d61f215809e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 853.679221] env[62600]: DEBUG oslo_concurrency.lockutils [req-63aa84ca-5792-43ee-9bfe-f52d6e77598a req-976a2721-9340-45fc-9bf2-0e60c6210577 service nova] Acquired lock "refresh_cache-d66c404d-8fd1-4fb7-a3b9-f21854c7e735" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.679426] env[62600]: DEBUG nova.network.neutron [req-63aa84ca-5792-43ee-9bfe-f52d6e77598a req-976a2721-9340-45fc-9bf2-0e60c6210577 service nova] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Refreshing network info cache for port 57606b61-13bc-4562-b28d-d61f215809e0 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 853.680996] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:05:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3e05affa-2640-435e-a124-0ee8a6ab1152', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '57606b61-13bc-4562-b28d-d61f215809e0', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 853.697063] env[62600]: DEBUG oslo.service.loopingcall [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.701867] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 853.702547] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d350b719-e6c9-403d-887f-986d60c53ada {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.729109] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 853.729109] env[62600]: value = "task-1222542" [ 853.729109] env[62600]: _type = "Task" [ 853.729109] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.741906] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222542, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.779246] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Releasing lock "refresh_cache-af7036df-b9f0-4ce6-962a-1edd7c1ea211" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.068353] env[62600]: DEBUG nova.scheduler.client.report [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 854.209701] env[62600]: DEBUG nova.network.neutron [req-63aa84ca-5792-43ee-9bfe-f52d6e77598a req-976a2721-9340-45fc-9bf2-0e60c6210577 service nova] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Updated VIF entry in instance network info cache for port 57606b61-13bc-4562-b28d-d61f215809e0. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 854.210235] env[62600]: DEBUG nova.network.neutron [req-63aa84ca-5792-43ee-9bfe-f52d6e77598a req-976a2721-9340-45fc-9bf2-0e60c6210577 service nova] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Updating instance_info_cache with network_info: [{"id": "57606b61-13bc-4562-b28d-d61f215809e0", "address": "fa:16:3e:53:05:09", "network": {"id": "036dc114-6ea0-46cb-a958-4a6137e7e85f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-728627542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e1772c9a6c4671ab89df2e8b40cae1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57606b61-13", "ovs_interfaceid": "57606b61-13bc-4562-b28d-d61f215809e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.240573] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222542, 'name': CreateVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.277248] env[62600]: DEBUG nova.network.neutron [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 854.474868] env[62600]: DEBUG nova.network.neutron [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.575474] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.400s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.576163] env[62600]: DEBUG nova.compute.manager [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 854.578891] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.701s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.579158] env[62600]: DEBUG nova.objects.instance [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Lazy-loading 'resources' on Instance uuid e25c631d-3d1b-40d2-9fb1-a65431f991a9 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 854.713543] env[62600]: DEBUG oslo_concurrency.lockutils [req-63aa84ca-5792-43ee-9bfe-f52d6e77598a req-976a2721-9340-45fc-9bf2-0e60c6210577 service nova] Releasing lock "refresh_cache-d66c404d-8fd1-4fb7-a3b9-f21854c7e735" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.745355] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222542, 'name': CreateVM_Task, 'duration_secs': 0.617932} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.745355] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 854.746044] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.746264] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.746679] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 854.747028] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ca277a2-f8aa-4b9e-9afe-6c3a358fe5ea {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.754999] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 854.754999] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52361041-1167-58bc-3740-84ced50500ee" [ 854.754999] env[62600]: _type = "Task" [ 854.754999] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.764267] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52361041-1167-58bc-3740-84ced50500ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.928080] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 854.929092] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-808737f3-37ae-40db-9ed0-94a4b31f5534 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.939957] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 854.941010] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0afcbbba-49af-4c32-a0a4-7b94df4bd0ed {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.980080] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Releasing lock "refresh_cache-ee7175b9-dbe6-4f90-bd2b-8829194dc6c3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.980080] env[62600]: DEBUG nova.compute.manager [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 854.980452] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 854.981244] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a702c577-d78f-43c1-9b4f-e154f446d4a5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.997950] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 855.000152] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2eff3a6d-15ba-4c78-8d39-574b39d7552c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.008878] env[62600]: DEBUG oslo_vmware.api [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for the task: (returnval){ [ 855.008878] env[62600]: value = "task-1222544" [ 855.008878] env[62600]: _type = "Task" [ 855.008878] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.021239] env[62600]: DEBUG oslo_vmware.api [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222544, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.022846] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 855.023116] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 855.023422] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Deleting the datastore file [datastore2] af7036df-b9f0-4ce6-962a-1edd7c1ea211 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 855.023799] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-77e158e4-db41-49f3-a7c7-b70d4bdd1011 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.033585] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 855.033585] env[62600]: value = "task-1222545" [ 855.033585] env[62600]: _type = "Task" [ 855.033585] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.043386] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222545, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.082164] env[62600]: DEBUG nova.compute.utils [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 855.083697] env[62600]: DEBUG nova.compute.manager [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 855.083875] env[62600]: DEBUG nova.network.neutron [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 855.096756] env[62600]: DEBUG nova.compute.manager [req-f1d71222-c96d-4c20-92b7-b33f76b86159 req-de1db173-4f89-4782-a66f-1080a11d03c1 service nova] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Received event network-vif-unplugged-b15ca3aa-0a63-4385-9171-98748d4b988c {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.097160] env[62600]: DEBUG oslo_concurrency.lockutils [req-f1d71222-c96d-4c20-92b7-b33f76b86159 req-de1db173-4f89-4782-a66f-1080a11d03c1 service nova] Acquiring lock "af7036df-b9f0-4ce6-962a-1edd7c1ea211-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.097438] env[62600]: DEBUG oslo_concurrency.lockutils [req-f1d71222-c96d-4c20-92b7-b33f76b86159 req-de1db173-4f89-4782-a66f-1080a11d03c1 service nova] Lock "af7036df-b9f0-4ce6-962a-1edd7c1ea211-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.099074] env[62600]: DEBUG oslo_concurrency.lockutils [req-f1d71222-c96d-4c20-92b7-b33f76b86159 req-de1db173-4f89-4782-a66f-1080a11d03c1 service nova] Lock "af7036df-b9f0-4ce6-962a-1edd7c1ea211-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.099074] env[62600]: DEBUG nova.compute.manager [req-f1d71222-c96d-4c20-92b7-b33f76b86159 req-de1db173-4f89-4782-a66f-1080a11d03c1 service nova] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] No waiting events found dispatching network-vif-unplugged-b15ca3aa-0a63-4385-9171-98748d4b988c {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 855.099074] env[62600]: WARNING nova.compute.manager [req-f1d71222-c96d-4c20-92b7-b33f76b86159 req-de1db173-4f89-4782-a66f-1080a11d03c1 service nova] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Received unexpected event network-vif-unplugged-b15ca3aa-0a63-4385-9171-98748d4b988c for instance with vm_state shelved and task_state shelving_offloading. [ 855.240306] env[62600]: DEBUG nova.policy [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8e824d803d54418aa119c4c9f36bb9f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7fc26d85259f46af8a82729fc20c6c60', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 855.267843] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52361041-1167-58bc-3740-84ced50500ee, 'name': SearchDatastore_Task, 'duration_secs': 0.013443} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.269032] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.269131] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 855.269428] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.270513] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.270513] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 855.270513] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a65e4741-4803-4aac-8353-266d6ed64e31 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.288962] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 855.289251] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 855.290407] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5279540-929a-4e1a-a20f-65c784ccbd83 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.301411] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 855.301411] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]523800b9-bece-0ac0-fbb0-d9138bd22493" [ 855.301411] env[62600]: _type = "Task" [ 855.301411] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.313085] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]523800b9-bece-0ac0-fbb0-d9138bd22493, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.447140] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5137fdfa-44d4-488a-a698-d42f02dadcc6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.461870] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5953ce17-0a27-45ef-9cf4-0e18992b353d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.510026] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4639fe18-06b4-4708-bc44-7b312ce48936 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.522294] env[62600]: DEBUG oslo_vmware.api [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222544, 'name': PowerOffVM_Task, 'duration_secs': 0.278321} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.523687] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb0675c-3a11-4ad7-b448-60cf87da59f6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.527613] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 855.527756] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 855.529587] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-19b641e8-a98d-450a-b3f1-b35929033f22 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.549245] env[62600]: DEBUG nova.compute.provider_tree [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.561611] env[62600]: DEBUG oslo_vmware.api [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222545, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151229} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.561856] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 855.562060] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 855.563117] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 855.566263] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 855.566482] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 855.566665] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Deleting the datastore file [datastore2] ee7175b9-dbe6-4f90-bd2b-8829194dc6c3 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 855.570390] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-63feea23-947f-4aa6-9a16-6d7f9557b2ea {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.577549] env[62600]: DEBUG oslo_vmware.api [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for the task: (returnval){ [ 855.577549] env[62600]: value = "task-1222547" [ 855.577549] env[62600]: _type = "Task" [ 855.577549] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.594644] env[62600]: DEBUG oslo_vmware.api [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222547, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.595839] env[62600]: INFO nova.scheduler.client.report [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Deleted allocations for instance af7036df-b9f0-4ce6-962a-1edd7c1ea211 [ 855.598688] env[62600]: DEBUG nova.compute.manager [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 855.624304] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Volume attach. Driver type: vmdk {{(pid=62600) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 855.624692] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264293', 'volume_id': '11e471b3-6f58-476c-a94e-4572bdbb8269', 'name': 'volume-11e471b3-6f58-476c-a94e-4572bdbb8269', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'afe838cc-e086-4986-87ec-4e1266bcaf60', 'attached_at': '', 'detached_at': '', 'volume_id': '11e471b3-6f58-476c-a94e-4572bdbb8269', 'serial': '11e471b3-6f58-476c-a94e-4572bdbb8269'} {{(pid=62600) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 855.625734] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-343d7ad9-016f-4ce4-b851-8cd2db6e8b86 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.662685] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1be161b1-78b9-4819-b4c3-b7bdcd083ce0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.697666] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] volume-11e471b3-6f58-476c-a94e-4572bdbb8269/volume-11e471b3-6f58-476c-a94e-4572bdbb8269.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 855.699262] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6224e95d-e5f3-41fe-b824-4513ae9c745c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.719362] env[62600]: DEBUG oslo_vmware.api [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 855.719362] env[62600]: value = "task-1222548" [ 855.719362] env[62600]: _type = "Task" [ 855.719362] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.730961] env[62600]: DEBUG oslo_vmware.api [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222548, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.814499] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]523800b9-bece-0ac0-fbb0-d9138bd22493, 'name': SearchDatastore_Task, 'duration_secs': 0.024655} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.815424] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f31de1d-98be-4e0d-a69e-af7d9338b2e4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.821397] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 855.821397] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520916db-5362-3cd8-1292-4b56d91bc909" [ 855.821397] env[62600]: _type = "Task" [ 855.821397] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.833249] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520916db-5362-3cd8-1292-4b56d91bc909, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.836384] env[62600]: DEBUG nova.network.neutron [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Successfully created port: c493ab77-1118-42b6-8009-41c5eced4543 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 856.052968] env[62600]: DEBUG nova.scheduler.client.report [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 856.090501] env[62600]: DEBUG oslo_vmware.api [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Task: {'id': task-1222547, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.102482} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.091235] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 856.091572] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 856.091886] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 856.092193] env[62600]: INFO nova.compute.manager [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Took 1.11 seconds to destroy the instance on the hypervisor. [ 856.092721] env[62600]: DEBUG oslo.service.loopingcall [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 856.094017] env[62600]: DEBUG nova.compute.manager [-] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 856.094017] env[62600]: DEBUG nova.network.neutron [-] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 856.105475] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.123053] env[62600]: DEBUG nova.network.neutron [-] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 856.234938] env[62600]: DEBUG oslo_vmware.api [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222548, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.335927] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520916db-5362-3cd8-1292-4b56d91bc909, 'name': SearchDatastore_Task, 'duration_secs': 0.034462} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.338013] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.338013] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] d66c404d-8fd1-4fb7-a3b9-f21854c7e735/d66c404d-8fd1-4fb7-a3b9-f21854c7e735.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 856.338013] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-94b248b7-ca57-4eae-8978-2c062600c00b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.344613] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 856.344613] env[62600]: value = "task-1222549" [ 856.344613] env[62600]: _type = "Task" [ 856.344613] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.354332] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222549, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.559517] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.980s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.562572] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.020s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.562881] env[62600]: DEBUG nova.objects.instance [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Lazy-loading 'resources' on Instance uuid 4e1a376f-6619-4c35-b75f-b45db75815cc {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.585440] env[62600]: INFO nova.scheduler.client.report [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Deleted allocations for instance e25c631d-3d1b-40d2-9fb1-a65431f991a9 [ 856.613315] env[62600]: DEBUG nova.compute.manager [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 856.624459] env[62600]: DEBUG nova.network.neutron [-] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.656048] env[62600]: DEBUG nova.virt.hardware [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 856.656427] env[62600]: DEBUG nova.virt.hardware [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 856.656611] env[62600]: DEBUG nova.virt.hardware [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 856.656812] env[62600]: DEBUG nova.virt.hardware [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 856.656996] env[62600]: DEBUG nova.virt.hardware [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 856.658109] env[62600]: DEBUG nova.virt.hardware [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 856.658526] env[62600]: DEBUG nova.virt.hardware [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 856.658731] env[62600]: DEBUG nova.virt.hardware [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 856.658947] env[62600]: DEBUG nova.virt.hardware [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 856.659157] env[62600]: DEBUG nova.virt.hardware [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 856.659410] env[62600]: DEBUG nova.virt.hardware [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 856.660937] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b76de3be-96db-4c14-a2c8-a81552aac022 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.675867] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d3ffc21-f051-4363-9086-78abcaf022b8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.734134] env[62600]: DEBUG oslo_vmware.api [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222548, 'name': ReconfigVM_Task, 'duration_secs': 0.604858} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.734599] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Reconfigured VM instance instance-00000037 to attach disk [datastore1] volume-11e471b3-6f58-476c-a94e-4572bdbb8269/volume-11e471b3-6f58-476c-a94e-4572bdbb8269.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 856.743176] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aaf92286-a123-4453-b65c-8250b13b3dec {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.764332] env[62600]: DEBUG oslo_vmware.api [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 856.764332] env[62600]: value = "task-1222550" [ 856.764332] env[62600]: _type = "Task" [ 856.764332] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.781728] env[62600]: DEBUG oslo_vmware.api [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222550, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.856776] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222549, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.860575] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0a69bddc-7b26-47dd-b61b-af438cda10d1 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "af7036df-b9f0-4ce6-962a-1edd7c1ea211" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.097667] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2c8d6bc2-2248-401e-9c5b-004ce8f9ad0b tempest-ServersTestFqdnHostnames-462438436 tempest-ServersTestFqdnHostnames-462438436-project-member] Lock "e25c631d-3d1b-40d2-9fb1-a65431f991a9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.789s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.129592] env[62600]: INFO nova.compute.manager [-] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Took 1.04 seconds to deallocate network for instance. [ 857.223692] env[62600]: DEBUG nova.compute.manager [req-b97da196-8acb-4cf4-a5b8-2f5cf3d243d8 req-dfc50766-9264-4f0c-93c7-70e7b3c4fd30 service nova] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Received event network-changed-b15ca3aa-0a63-4385-9171-98748d4b988c {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.223692] env[62600]: DEBUG nova.compute.manager [req-b97da196-8acb-4cf4-a5b8-2f5cf3d243d8 req-dfc50766-9264-4f0c-93c7-70e7b3c4fd30 service nova] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Refreshing instance network info cache due to event network-changed-b15ca3aa-0a63-4385-9171-98748d4b988c. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 857.223692] env[62600]: DEBUG oslo_concurrency.lockutils [req-b97da196-8acb-4cf4-a5b8-2f5cf3d243d8 req-dfc50766-9264-4f0c-93c7-70e7b3c4fd30 service nova] Acquiring lock "refresh_cache-af7036df-b9f0-4ce6-962a-1edd7c1ea211" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.223692] env[62600]: DEBUG oslo_concurrency.lockutils [req-b97da196-8acb-4cf4-a5b8-2f5cf3d243d8 req-dfc50766-9264-4f0c-93c7-70e7b3c4fd30 service nova] Acquired lock "refresh_cache-af7036df-b9f0-4ce6-962a-1edd7c1ea211" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.223904] env[62600]: DEBUG nova.network.neutron [req-b97da196-8acb-4cf4-a5b8-2f5cf3d243d8 req-dfc50766-9264-4f0c-93c7-70e7b3c4fd30 service nova] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Refreshing network info cache for port b15ca3aa-0a63-4385-9171-98748d4b988c {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 857.277623] env[62600]: DEBUG oslo_vmware.api [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222550, 'name': ReconfigVM_Task, 'duration_secs': 0.241454} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.277941] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264293', 'volume_id': '11e471b3-6f58-476c-a94e-4572bdbb8269', 'name': 'volume-11e471b3-6f58-476c-a94e-4572bdbb8269', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'afe838cc-e086-4986-87ec-4e1266bcaf60', 'attached_at': '', 'detached_at': '', 'volume_id': '11e471b3-6f58-476c-a94e-4572bdbb8269', 'serial': '11e471b3-6f58-476c-a94e-4572bdbb8269'} {{(pid=62600) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 857.356221] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf448239-06b4-4db3-86bb-4de146ccdead {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.367048] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222549, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.625545} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.369231] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] d66c404d-8fd1-4fb7-a3b9-f21854c7e735/d66c404d-8fd1-4fb7-a3b9-f21854c7e735.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 857.369475] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 857.369790] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d3537168-10b6-4807-841a-833fd638fdbf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.372512] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d2b6e33-cf47-4e8c-b97f-769229c7a413 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.405603] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2c90b2f-f8f5-4843-b357-49f34de6fa68 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.408486] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 857.408486] env[62600]: value = "task-1222551" [ 857.408486] env[62600]: _type = "Task" [ 857.408486] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.417265] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2cb4604-190d-4944-ad17-2c342ae3c7b6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.424725] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222551, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.435177] env[62600]: DEBUG nova.compute.provider_tree [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.622136] env[62600]: DEBUG nova.network.neutron [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Successfully updated port: c493ab77-1118-42b6-8009-41c5eced4543 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.636527] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.764874] env[62600]: DEBUG oslo_vmware.rw_handles [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524febcb-0961-8ec1-6050-b89178972c4b/disk-0.vmdk. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 857.765726] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f49a5ff4-63b0-4953-81da-90ffe9c0b6cc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.772617] env[62600]: DEBUG oslo_vmware.rw_handles [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524febcb-0961-8ec1-6050-b89178972c4b/disk-0.vmdk is in state: ready. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 857.772800] env[62600]: ERROR oslo_vmware.rw_handles [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524febcb-0961-8ec1-6050-b89178972c4b/disk-0.vmdk due to incomplete transfer. [ 857.773052] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-cc154629-47ee-4981-abbb-352460c2fa5f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.782896] env[62600]: DEBUG oslo_vmware.rw_handles [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524febcb-0961-8ec1-6050-b89178972c4b/disk-0.vmdk. {{(pid=62600) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 857.783116] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Uploaded image e1395438-00bb-4e48-add8-ef3068f3c1c3 to the Glance image server {{(pid=62600) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 857.784716] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Destroying the VM {{(pid=62600) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 857.785305] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d1e78270-2354-4db8-8055-4cf24bc270ff {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.791065] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 857.791065] env[62600]: value = "task-1222552" [ 857.791065] env[62600]: _type = "Task" [ 857.791065] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.799514] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222552, 'name': Destroy_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.920316] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222551, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.124412} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.920575] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 857.921534] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c999e50c-01b2-4fb2-8f51-a4f8d13efd32 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.948317] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] d66c404d-8fd1-4fb7-a3b9-f21854c7e735/d66c404d-8fd1-4fb7-a3b9-f21854c7e735.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 857.949292] env[62600]: DEBUG nova.scheduler.client.report [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 857.952554] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-987d7365-f1c7-497f-bc42-d2e19291ddbc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.978454] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 857.978454] env[62600]: value = "task-1222553" [ 857.978454] env[62600]: _type = "Task" [ 857.978454] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.987503] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222553, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.125648] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Acquiring lock "refresh_cache-ddc80377-a14e-48d0-ad24-cc6daf7f6850" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.126924] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Acquired lock "refresh_cache-ddc80377-a14e-48d0-ad24-cc6daf7f6850" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.126924] env[62600]: DEBUG nova.network.neutron [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 858.302996] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222552, 'name': Destroy_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.314080] env[62600]: DEBUG nova.network.neutron [req-b97da196-8acb-4cf4-a5b8-2f5cf3d243d8 req-dfc50766-9264-4f0c-93c7-70e7b3c4fd30 service nova] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Updated VIF entry in instance network info cache for port b15ca3aa-0a63-4385-9171-98748d4b988c. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 858.314080] env[62600]: DEBUG nova.network.neutron [req-b97da196-8acb-4cf4-a5b8-2f5cf3d243d8 req-dfc50766-9264-4f0c-93c7-70e7b3c4fd30 service nova] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Updating instance_info_cache with network_info: [{"id": "b15ca3aa-0a63-4385-9171-98748d4b988c", "address": "fa:16:3e:0d:e9:b6", "network": {"id": "5ef67e8a-5613-4cdb-9f35-0a18720657ff", "bridge": null, "label": "tempest-DeleteServersTestJSON-1644754500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bab74e3ae78248909dbcd483abbb8da2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapb15ca3aa-0a", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.340415] env[62600]: DEBUG nova.objects.instance [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lazy-loading 'flavor' on Instance uuid afe838cc-e086-4986-87ec-4e1266bcaf60 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 858.470515] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.908s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.473131] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.609s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.476501] env[62600]: DEBUG nova.objects.instance [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Lazy-loading 'resources' on Instance uuid 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 858.490820] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222553, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.501552] env[62600]: INFO nova.scheduler.client.report [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Deleted allocations for instance 4e1a376f-6619-4c35-b75f-b45db75815cc [ 858.684946] env[62600]: DEBUG nova.network.neutron [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 858.802126] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222552, 'name': Destroy_Task, 'duration_secs': 0.835875} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.802557] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Destroyed the VM [ 858.803604] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Deleting Snapshot of the VM instance {{(pid=62600) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 858.804069] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5927513b-4239-41b6-9b53-9dc5ee621826 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.812023] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 858.812023] env[62600]: value = "task-1222554" [ 858.812023] env[62600]: _type = "Task" [ 858.812023] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.816374] env[62600]: DEBUG oslo_concurrency.lockutils [req-b97da196-8acb-4cf4-a5b8-2f5cf3d243d8 req-dfc50766-9264-4f0c-93c7-70e7b3c4fd30 service nova] Releasing lock "refresh_cache-af7036df-b9f0-4ce6-962a-1edd7c1ea211" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.821957] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222554, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.850279] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f4def227-ce98-48e7-9551-10526be8a098 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "afe838cc-e086-4986-87ec-4e1266bcaf60" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.905s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.948014] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c9adedce-3dee-4b29-b077-4d5a003a7335 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquiring lock "afe838cc-e086-4986-87ec-4e1266bcaf60" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.948014] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c9adedce-3dee-4b29-b077-4d5a003a7335 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "afe838cc-e086-4986-87ec-4e1266bcaf60" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.997017] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222553, 'name': ReconfigVM_Task, 'duration_secs': 0.749551} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.997017] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Reconfigured VM instance instance-00000042 to attach disk [datastore1] d66c404d-8fd1-4fb7-a3b9-f21854c7e735/d66c404d-8fd1-4fb7-a3b9-f21854c7e735.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 858.997017] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2c7912ed-fc95-4014-8247-c2bca0fbc1b3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.001597] env[62600]: DEBUG nova.network.neutron [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Updating instance_info_cache with network_info: [{"id": "c493ab77-1118-42b6-8009-41c5eced4543", "address": "fa:16:3e:b6:f2:fc", "network": {"id": "d309fe63-baa4-4f9a-b414-276eec07ee86", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1080590449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fc26d85259f46af8a82729fc20c6c60", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8480e3f5-68bd-4c27-ae1f-7c994a8202b1", "external-id": "nsx-vlan-transportzone-628", "segmentation_id": 628, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc493ab77-11", "ovs_interfaceid": "c493ab77-1118-42b6-8009-41c5eced4543", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.019412] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 859.019412] env[62600]: value = "task-1222555" [ 859.019412] env[62600]: _type = "Task" [ 859.019412] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.020209] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1e207cf7-b43a-4f29-b615-6280f6a6973b tempest-InstanceActionsV221TestJSON-197847864 tempest-InstanceActionsV221TestJSON-197847864-project-member] Lock "4e1a376f-6619-4c35-b75f-b45db75815cc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.263s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.034299] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222555, 'name': Rename_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.262655] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef982186-a63f-41bc-ab57-f2b40612b58d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.273433] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-357fa65f-6dd8-4ce0-9e1c-208270ee0dbf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.311525] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee3e54ae-8958-4bff-81d1-7e9befa6dc46 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.325520] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28861c04-d9f7-4240-b184-610c97ae7348 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.329613] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222554, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.339442] env[62600]: DEBUG nova.compute.provider_tree [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.450737] env[62600]: INFO nova.compute.manager [None req-c9adedce-3dee-4b29-b077-4d5a003a7335 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Detaching volume 11e471b3-6f58-476c-a94e-4572bdbb8269 [ 859.506390] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Releasing lock "refresh_cache-ddc80377-a14e-48d0-ad24-cc6daf7f6850" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.506779] env[62600]: DEBUG nova.compute.manager [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Instance network_info: |[{"id": "c493ab77-1118-42b6-8009-41c5eced4543", "address": "fa:16:3e:b6:f2:fc", "network": {"id": "d309fe63-baa4-4f9a-b414-276eec07ee86", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1080590449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fc26d85259f46af8a82729fc20c6c60", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8480e3f5-68bd-4c27-ae1f-7c994a8202b1", "external-id": "nsx-vlan-transportzone-628", "segmentation_id": 628, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc493ab77-11", "ovs_interfaceid": "c493ab77-1118-42b6-8009-41c5eced4543", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 859.508497] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:f2:fc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8480e3f5-68bd-4c27-ae1f-7c994a8202b1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c493ab77-1118-42b6-8009-41c5eced4543', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 859.517528] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Creating folder: Project (7fc26d85259f46af8a82729fc20c6c60). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 859.518863] env[62600]: INFO nova.virt.block_device [None req-c9adedce-3dee-4b29-b077-4d5a003a7335 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Attempting to driver detach volume 11e471b3-6f58-476c-a94e-4572bdbb8269 from mountpoint /dev/sdb [ 859.519095] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9adedce-3dee-4b29-b077-4d5a003a7335 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Volume detach. Driver type: vmdk {{(pid=62600) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 859.519297] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9adedce-3dee-4b29-b077-4d5a003a7335 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264293', 'volume_id': '11e471b3-6f58-476c-a94e-4572bdbb8269', 'name': 'volume-11e471b3-6f58-476c-a94e-4572bdbb8269', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'afe838cc-e086-4986-87ec-4e1266bcaf60', 'attached_at': '', 'detached_at': '', 'volume_id': '11e471b3-6f58-476c-a94e-4572bdbb8269', 'serial': '11e471b3-6f58-476c-a94e-4572bdbb8269'} {{(pid=62600) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 859.519574] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0ec1f44c-ebcf-4b74-8f70-577b11b285bf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.522646] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65878cf3-e470-4dcc-98d3-0baf31082579 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.537223] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222555, 'name': Rename_Task, 'duration_secs': 0.316535} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.554733] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 859.559020] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-581728c1-a470-44e3-a6e5-f78d2161c9c2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.563271] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c40d611-0064-4f9a-9de7-f866933bad8e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.564422] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Created folder: Project (7fc26d85259f46af8a82729fc20c6c60) in parent group-v264198. [ 859.564640] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Creating folder: Instances. Parent ref: group-v264295. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 859.564889] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-87840ea1-cb55-4d85-9d8b-29f15cd02174 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.572691] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb0afd2-b3aa-40d1-9b17-34b425537bca {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.575379] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 859.575379] env[62600]: value = "task-1222557" [ 859.575379] env[62600]: _type = "Task" [ 859.575379] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.577438] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Created folder: Instances in parent group-v264295. [ 859.577438] env[62600]: DEBUG oslo.service.loopingcall [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 859.580785] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 859.601816] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a687a699-4203-457f-a003-68a35070b786 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.619614] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56953c13-f898-42cc-a3ba-ca943f818185 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.628392] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222557, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.641907] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9adedce-3dee-4b29-b077-4d5a003a7335 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] The volume has not been displaced from its original location: [datastore1] volume-11e471b3-6f58-476c-a94e-4572bdbb8269/volume-11e471b3-6f58-476c-a94e-4572bdbb8269.vmdk. No consolidation needed. {{(pid=62600) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 859.647430] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9adedce-3dee-4b29-b077-4d5a003a7335 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Reconfiguring VM instance instance-00000037 to detach disk 2001 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 859.650313] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ace57d9d-9a65-4098-ae89-09ff35ff48c1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.664057] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 859.664057] env[62600]: value = "task-1222559" [ 859.664057] env[62600]: _type = "Task" [ 859.664057] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.667802] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.667891] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.676851] env[62600]: DEBUG oslo_vmware.api [None req-c9adedce-3dee-4b29-b077-4d5a003a7335 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 859.676851] env[62600]: value = "task-1222560" [ 859.676851] env[62600]: _type = "Task" [ 859.676851] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.682095] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222559, 'name': CreateVM_Task} progress is 15%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.692548] env[62600]: DEBUG oslo_vmware.api [None req-c9adedce-3dee-4b29-b077-4d5a003a7335 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222560, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.757701] env[62600]: DEBUG nova.compute.manager [req-5439f294-eea9-4037-8717-ec5a868b5f4b req-57a75cce-7f08-4c38-affc-854ac302fff8 service nova] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Received event network-vif-plugged-c493ab77-1118-42b6-8009-41c5eced4543 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 859.758084] env[62600]: DEBUG oslo_concurrency.lockutils [req-5439f294-eea9-4037-8717-ec5a868b5f4b req-57a75cce-7f08-4c38-affc-854ac302fff8 service nova] Acquiring lock "ddc80377-a14e-48d0-ad24-cc6daf7f6850-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.758432] env[62600]: DEBUG oslo_concurrency.lockutils [req-5439f294-eea9-4037-8717-ec5a868b5f4b req-57a75cce-7f08-4c38-affc-854ac302fff8 service nova] Lock "ddc80377-a14e-48d0-ad24-cc6daf7f6850-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.758809] env[62600]: DEBUG oslo_concurrency.lockutils [req-5439f294-eea9-4037-8717-ec5a868b5f4b req-57a75cce-7f08-4c38-affc-854ac302fff8 service nova] Lock "ddc80377-a14e-48d0-ad24-cc6daf7f6850-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.759049] env[62600]: DEBUG nova.compute.manager [req-5439f294-eea9-4037-8717-ec5a868b5f4b req-57a75cce-7f08-4c38-affc-854ac302fff8 service nova] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] No waiting events found dispatching network-vif-plugged-c493ab77-1118-42b6-8009-41c5eced4543 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 859.759582] env[62600]: WARNING nova.compute.manager [req-5439f294-eea9-4037-8717-ec5a868b5f4b req-57a75cce-7f08-4c38-affc-854ac302fff8 service nova] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Received unexpected event network-vif-plugged-c493ab77-1118-42b6-8009-41c5eced4543 for instance with vm_state building and task_state spawning. [ 859.759660] env[62600]: DEBUG nova.compute.manager [req-5439f294-eea9-4037-8717-ec5a868b5f4b req-57a75cce-7f08-4c38-affc-854ac302fff8 service nova] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Received event network-changed-c493ab77-1118-42b6-8009-41c5eced4543 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 859.759903] env[62600]: DEBUG nova.compute.manager [req-5439f294-eea9-4037-8717-ec5a868b5f4b req-57a75cce-7f08-4c38-affc-854ac302fff8 service nova] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Refreshing instance network info cache due to event network-changed-c493ab77-1118-42b6-8009-41c5eced4543. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 859.760295] env[62600]: DEBUG oslo_concurrency.lockutils [req-5439f294-eea9-4037-8717-ec5a868b5f4b req-57a75cce-7f08-4c38-affc-854ac302fff8 service nova] Acquiring lock "refresh_cache-ddc80377-a14e-48d0-ad24-cc6daf7f6850" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.760549] env[62600]: DEBUG oslo_concurrency.lockutils [req-5439f294-eea9-4037-8717-ec5a868b5f4b req-57a75cce-7f08-4c38-affc-854ac302fff8 service nova] Acquired lock "refresh_cache-ddc80377-a14e-48d0-ad24-cc6daf7f6850" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.760842] env[62600]: DEBUG nova.network.neutron [req-5439f294-eea9-4037-8717-ec5a868b5f4b req-57a75cce-7f08-4c38-affc-854ac302fff8 service nova] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Refreshing network info cache for port c493ab77-1118-42b6-8009-41c5eced4543 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 859.825031] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222554, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.842917] env[62600]: DEBUG nova.scheduler.client.report [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 860.090548] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222557, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.173509] env[62600]: DEBUG nova.compute.manager [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 860.187247] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222559, 'name': CreateVM_Task, 'duration_secs': 0.543387} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.200039] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 860.200039] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.200039] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.200039] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 860.200039] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80674c2d-e9e0-4636-ad2b-76a4e6ad9f06 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.211732] env[62600]: DEBUG oslo_vmware.api [None req-c9adedce-3dee-4b29-b077-4d5a003a7335 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222560, 'name': ReconfigVM_Task, 'duration_secs': 0.338212} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.216183] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9adedce-3dee-4b29-b077-4d5a003a7335 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Reconfigured VM instance instance-00000037 to detach disk 2001 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 860.223129] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1985fdb1-0af8-4033-93fd-682a90b0cd38 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.233387] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Waiting for the task: (returnval){ [ 860.233387] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]523bed83-7ebc-17b8-3b73-1fb4974e5418" [ 860.233387] env[62600]: _type = "Task" [ 860.233387] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.240735] env[62600]: DEBUG oslo_vmware.api [None req-c9adedce-3dee-4b29-b077-4d5a003a7335 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 860.240735] env[62600]: value = "task-1222561" [ 860.240735] env[62600]: _type = "Task" [ 860.240735] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.249398] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]523bed83-7ebc-17b8-3b73-1fb4974e5418, 'name': SearchDatastore_Task, 'duration_secs': 0.018224} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.250361] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.250633] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.250873] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.251024] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.251557] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.255341] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-53750ce3-db71-48f4-b105-cc02328cacb8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.257369] env[62600]: DEBUG oslo_vmware.api [None req-c9adedce-3dee-4b29-b077-4d5a003a7335 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222561, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.279609] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.279916] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 860.280677] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9aa00a9-0d9f-4e8f-815d-610bb49081ed {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.288013] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Waiting for the task: (returnval){ [ 860.288013] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b631f6-5e0c-4d38-0665-45e25ad1233a" [ 860.288013] env[62600]: _type = "Task" [ 860.288013] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.296220] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b631f6-5e0c-4d38-0665-45e25ad1233a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.325927] env[62600]: DEBUG oslo_vmware.api [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222554, 'name': RemoveSnapshot_Task, 'duration_secs': 1.081976} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.329768] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Deleted Snapshot of the VM instance {{(pid=62600) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 860.330754] env[62600]: INFO nova.compute.manager [None req-94d47086-3f3b-4368-9923-7d6786fb24c2 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Took 18.83 seconds to snapshot the instance on the hypervisor. [ 860.349015] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.876s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.351661] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.102s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.352118] env[62600]: DEBUG nova.objects.instance [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lazy-loading 'resources' on Instance uuid b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 860.379980] env[62600]: INFO nova.scheduler.client.report [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Deleted allocations for instance 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4 [ 860.592961] env[62600]: DEBUG oslo_vmware.api [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222557, 'name': PowerOnVM_Task, 'duration_secs': 0.714595} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.592961] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 860.592961] env[62600]: INFO nova.compute.manager [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Took 10.39 seconds to spawn the instance on the hypervisor. [ 860.592961] env[62600]: DEBUG nova.compute.manager [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 860.593303] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85aa4102-4455-4cb0-9770-c66cc6c761ac {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.711437] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.730623] env[62600]: DEBUG nova.network.neutron [req-5439f294-eea9-4037-8717-ec5a868b5f4b req-57a75cce-7f08-4c38-affc-854ac302fff8 service nova] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Updated VIF entry in instance network info cache for port c493ab77-1118-42b6-8009-41c5eced4543. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 860.730623] env[62600]: DEBUG nova.network.neutron [req-5439f294-eea9-4037-8717-ec5a868b5f4b req-57a75cce-7f08-4c38-affc-854ac302fff8 service nova] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Updating instance_info_cache with network_info: [{"id": "c493ab77-1118-42b6-8009-41c5eced4543", "address": "fa:16:3e:b6:f2:fc", "network": {"id": "d309fe63-baa4-4f9a-b414-276eec07ee86", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1080590449-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7fc26d85259f46af8a82729fc20c6c60", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8480e3f5-68bd-4c27-ae1f-7c994a8202b1", "external-id": "nsx-vlan-transportzone-628", "segmentation_id": 628, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc493ab77-11", "ovs_interfaceid": "c493ab77-1118-42b6-8009-41c5eced4543", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.755211] env[62600]: DEBUG oslo_vmware.api [None req-c9adedce-3dee-4b29-b077-4d5a003a7335 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222561, 'name': ReconfigVM_Task, 'duration_secs': 0.206353} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.755211] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9adedce-3dee-4b29-b077-4d5a003a7335 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264293', 'volume_id': '11e471b3-6f58-476c-a94e-4572bdbb8269', 'name': 'volume-11e471b3-6f58-476c-a94e-4572bdbb8269', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'afe838cc-e086-4986-87ec-4e1266bcaf60', 'attached_at': '', 'detached_at': '', 'volume_id': '11e471b3-6f58-476c-a94e-4572bdbb8269', 'serial': '11e471b3-6f58-476c-a94e-4572bdbb8269'} {{(pid=62600) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 860.804020] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b631f6-5e0c-4d38-0665-45e25ad1233a, 'name': SearchDatastore_Task, 'duration_secs': 0.028714} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.804020] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d2efd6e-9f92-4a35-9c6e-acddc6630a2a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.810667] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Waiting for the task: (returnval){ [ 860.810667] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526d5994-0a5f-350f-c56f-30899e847121" [ 860.810667] env[62600]: _type = "Task" [ 860.810667] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.822396] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526d5994-0a5f-350f-c56f-30899e847121, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.895496] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7fb813a0-d42c-4adc-8132-6210ec0c5407 tempest-ServersTestJSON-1739708439 tempest-ServersTestJSON-1739708439-project-member] Lock "0a8840f1-7681-4aaa-9f3a-32b72c04c1c4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.461s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.971913] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "ab11cdd1-80a7-43c4-afac-24c7930238c1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.972321] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "ab11cdd1-80a7-43c4-afac-24c7930238c1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.121328] env[62600]: INFO nova.compute.manager [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Took 46.67 seconds to build instance. [ 861.171993] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f98157e4-3ded-46fc-b78f-c8b05eaa578d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.179666] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4afd7507-8eea-4295-8186-334f6beb6e1f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.214204] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a33531f-3600-45f3-a9ea-e5688782e8e0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.222109] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3366e524-e064-4636-bf1a-24153fbef9b2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.237509] env[62600]: DEBUG oslo_concurrency.lockutils [req-5439f294-eea9-4037-8717-ec5a868b5f4b req-57a75cce-7f08-4c38-affc-854ac302fff8 service nova] Releasing lock "refresh_cache-ddc80377-a14e-48d0-ad24-cc6daf7f6850" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.239701] env[62600]: DEBUG nova.compute.provider_tree [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.322517] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526d5994-0a5f-350f-c56f-30899e847121, 'name': SearchDatastore_Task, 'duration_secs': 0.0121} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.322910] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.323279] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] ddc80377-a14e-48d0-ad24-cc6daf7f6850/ddc80377-a14e-48d0-ad24-cc6daf7f6850.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 861.323607] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eb45ecba-9a7d-476e-916b-f18b21e05a44 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.332118] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Waiting for the task: (returnval){ [ 861.332118] env[62600]: value = "task-1222562" [ 861.332118] env[62600]: _type = "Task" [ 861.332118] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.339243] env[62600]: DEBUG nova.objects.instance [None req-c9adedce-3dee-4b29-b077-4d5a003a7335 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lazy-loading 'flavor' on Instance uuid afe838cc-e086-4986-87ec-4e1266bcaf60 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 861.346747] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': task-1222562, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.476209] env[62600]: DEBUG nova.compute.manager [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 861.627653] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e5c018ad-85bd-4c86-9bdc-682d68bb4436 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "d66c404d-8fd1-4fb7-a3b9-f21854c7e735" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.561s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.745058] env[62600]: DEBUG nova.scheduler.client.report [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 861.753709] env[62600]: INFO nova.compute.manager [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Rescuing [ 861.753965] env[62600]: DEBUG oslo_concurrency.lockutils [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "refresh_cache-d66c404d-8fd1-4fb7-a3b9-f21854c7e735" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.754130] env[62600]: DEBUG oslo_concurrency.lockutils [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquired lock "refresh_cache-d66c404d-8fd1-4fb7-a3b9-f21854c7e735" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.754928] env[62600]: DEBUG nova.network.neutron [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 861.848107] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': task-1222562, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.999871] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.257390] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.906s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.265898] env[62600]: DEBUG oslo_concurrency.lockutils [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.539s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.266215] env[62600]: DEBUG nova.objects.instance [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Lazy-loading 'resources' on Instance uuid fb6f8e36-8d24-45ea-a6e4-4d768c3b232d {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.272129] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Acquiring lock "f6957b5a-5da8-4205-92e5-2f08bf948e88" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.272333] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Lock "f6957b5a-5da8-4205-92e5-2f08bf948e88" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.289873] env[62600]: INFO nova.scheduler.client.report [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Deleted allocations for instance b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc [ 862.350873] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': task-1222562, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.351343] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c9adedce-3dee-4b29-b077-4d5a003a7335 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "afe838cc-e086-4986-87ec-4e1266bcaf60" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.404s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.634136] env[62600]: DEBUG nova.network.neutron [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Updating instance_info_cache with network_info: [{"id": "57606b61-13bc-4562-b28d-d61f215809e0", "address": "fa:16:3e:53:05:09", "network": {"id": "036dc114-6ea0-46cb-a958-4a6137e7e85f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-728627542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e1772c9a6c4671ab89df2e8b40cae1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57606b61-13", "ovs_interfaceid": "57606b61-13bc-4562-b28d-d61f215809e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.781720] env[62600]: DEBUG nova.compute.manager [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 862.812590] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e9d754a0-9d5a-4878-84ae-ec9dcdf54c9b tempest-ServerShowV247Test-323146770 tempest-ServerShowV247Test-323146770-project-member] Lock "b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.320s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.858331] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': task-1222562, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.968016] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "interface-bd9b7351-01a2-429d-a860-d9ff6855eefc-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.968517] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "interface-bd9b7351-01a2-429d-a860-d9ff6855eefc-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.969074] env[62600]: DEBUG nova.objects.instance [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lazy-loading 'flavor' on Instance uuid bd9b7351-01a2-429d-a860-d9ff6855eefc {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 863.097802] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd940cb8-902a-436a-b090-0d6f0d277af7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.108532] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7096ade0-5c98-4544-8a42-6761b7c37843 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.142373] env[62600]: DEBUG oslo_concurrency.lockutils [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Releasing lock "refresh_cache-d66c404d-8fd1-4fb7-a3b9-f21854c7e735" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.145486] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dccae12b-7b00-4c9d-ad86-60740c1f73d9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.156855] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5156c72e-0aef-4761-8874-b78a895ed6e4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.175468] env[62600]: DEBUG nova.compute.provider_tree [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.310265] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.345937] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': task-1222562, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.642632} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.346295] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] ddc80377-a14e-48d0-ad24-cc6daf7f6850/ddc80377-a14e-48d0-ad24-cc6daf7f6850.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 863.346747] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 863.346886] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18c8e777-b1c0-4247-ad8b-7f87b881fcee {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.355819] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Waiting for the task: (returnval){ [ 863.355819] env[62600]: value = "task-1222563" [ 863.355819] env[62600]: _type = "Task" [ 863.355819] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.365490] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': task-1222563, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.674255] env[62600]: DEBUG nova.objects.instance [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lazy-loading 'pci_requests' on Instance uuid bd9b7351-01a2-429d-a860-d9ff6855eefc {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 863.686021] env[62600]: DEBUG nova.scheduler.client.report [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 863.690334] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 863.691460] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ac60f3db-9f45-4926-b7ea-4bad341ccde4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.703294] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 863.703294] env[62600]: value = "task-1222564" [ 863.703294] env[62600]: _type = "Task" [ 863.703294] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.719424] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222564, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.866672] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': task-1222563, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063539} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.866960] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 863.869688] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d0006e-9850-4919-b670-11348bb9d70f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.894541] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] ddc80377-a14e-48d0-ad24-cc6daf7f6850/ddc80377-a14e-48d0-ad24-cc6daf7f6850.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 863.895208] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-354897bc-0ecc-4267-a76e-53bca372a1f5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.916147] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Waiting for the task: (returnval){ [ 863.916147] env[62600]: value = "task-1222565" [ 863.916147] env[62600]: _type = "Task" [ 863.916147] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.924877] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': task-1222565, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.179183] env[62600]: DEBUG nova.objects.base [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62600) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 864.179538] env[62600]: DEBUG nova.network.neutron [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 864.192470] env[62600]: DEBUG oslo_concurrency.lockutils [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.927s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.195274] env[62600]: DEBUG oslo_concurrency.lockutils [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 21.284s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.195663] env[62600]: DEBUG nova.objects.instance [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62600) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 864.214614] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222564, 'name': PowerOffVM_Task, 'duration_secs': 0.203896} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.215080] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 864.215971] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-458079ac-1da0-4d94-9305-8939871c614e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.224020] env[62600]: INFO nova.scheduler.client.report [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Deleted allocations for instance fb6f8e36-8d24-45ea-a6e4-4d768c3b232d [ 864.248367] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e903a8e1-3c5a-4082-b834-8ea522d6568d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.253931] env[62600]: DEBUG nova.policy [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fbd1b2fb34d841359ada8fc44bec2986', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b028450f2da445fb83e37adfc86bba68', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 864.301684] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 864.302063] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e387ea8-40ac-4812-be76-c16432e8884c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.310033] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 864.310033] env[62600]: value = "task-1222566" [ 864.310033] env[62600]: _type = "Task" [ 864.310033] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.332286] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] VM already powered off {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 864.332855] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 864.333263] env[62600]: DEBUG oslo_concurrency.lockutils [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.333773] env[62600]: DEBUG oslo_concurrency.lockutils [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.334116] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 864.335444] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3ef91330-377e-4b6a-8ab4-a11e0910c0f1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.351756] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquiring lock "107ed783-10a3-4f07-a809-a5a076c2b904" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.352235] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "107ed783-10a3-4f07-a809-a5a076c2b904" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.355320] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 864.355320] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 864.356044] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-193106ab-808b-4077-9ab7-19d17bf8d1f6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.362745] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 864.362745] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d32397-f05f-530f-d0a3-c9ba0fe5800b" [ 864.362745] env[62600]: _type = "Task" [ 864.362745] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.374333] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d32397-f05f-530f-d0a3-c9ba0fe5800b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.428030] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': task-1222565, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.752323] env[62600]: DEBUG oslo_concurrency.lockutils [None req-167cf1d6-aafb-498d-b441-a32d28088a3c tempest-ServersNegativeTestMultiTenantJSON-1384572939 tempest-ServersNegativeTestMultiTenantJSON-1384572939-project-member] Lock "fb6f8e36-8d24-45ea-a6e4-4d768c3b232d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.021s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.857090] env[62600]: DEBUG nova.compute.manager [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 864.881576] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d32397-f05f-530f-d0a3-c9ba0fe5800b, 'name': SearchDatastore_Task, 'duration_secs': 0.023255} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.883032] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2c138cd-86d7-4e53-9cc7-c2646a3de75e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.893667] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 864.893667] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52dc6a52-b043-6cc1-6925-ecf114fa2730" [ 864.893667] env[62600]: _type = "Task" [ 864.893667] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.903869] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52dc6a52-b043-6cc1-6925-ecf114fa2730, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.926317] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': task-1222565, 'name': ReconfigVM_Task, 'duration_secs': 0.905878} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.926579] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Reconfigured VM instance instance-00000043 to attach disk [datastore1] ddc80377-a14e-48d0-ad24-cc6daf7f6850/ddc80377-a14e-48d0-ad24-cc6daf7f6850.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 864.927208] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-af624ad7-2b20-4ae5-a37f-b6d62efc9f4d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.929597] env[62600]: DEBUG nova.network.neutron [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Successfully created port: a0ecf848-49ae-489e-9c07-0acfa758d69f {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 864.937467] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Waiting for the task: (returnval){ [ 864.937467] env[62600]: value = "task-1222567" [ 864.937467] env[62600]: _type = "Task" [ 864.937467] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.948639] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': task-1222567, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.206587] env[62600]: DEBUG oslo_concurrency.lockutils [None req-02075e3c-9897-4491-ba51-c9ff9a5c7918 tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.211020] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.811s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.211020] env[62600]: DEBUG nova.objects.instance [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Lazy-loading 'resources' on Instance uuid c2c20e5f-6c0e-4989-bc36-ff4a27d5c534 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 865.398588] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.406414] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52dc6a52-b043-6cc1-6925-ecf114fa2730, 'name': SearchDatastore_Task, 'duration_secs': 0.010789} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.406932] env[62600]: DEBUG oslo_concurrency.lockutils [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.407578] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] d66c404d-8fd1-4fb7-a3b9-f21854c7e735/e7bd1d77-bfff-4684-9545-f6a1c69efe58-rescue.vmdk. {{(pid=62600) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 865.407877] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4e164ad8-2f89-44b5-b124-a6cb5bf99de2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.415260] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 865.415260] env[62600]: value = "task-1222568" [ 865.415260] env[62600]: _type = "Task" [ 865.415260] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.428093] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222568, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.446996] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': task-1222567, 'name': Rename_Task, 'duration_secs': 0.369357} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.447578] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 865.447929] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef95e764-6824-4e07-8d52-e825ed55dce9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.455223] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Waiting for the task: (returnval){ [ 865.455223] env[62600]: value = "task-1222569" [ 865.455223] env[62600]: _type = "Task" [ 865.455223] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.470757] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': task-1222569, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.931763] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222568, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.462419} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.932552] env[62600]: INFO nova.virt.vmwareapi.ds_util [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] d66c404d-8fd1-4fb7-a3b9-f21854c7e735/e7bd1d77-bfff-4684-9545-f6a1c69efe58-rescue.vmdk. [ 865.936821] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ffe6d23-695f-4eba-af62-a6ddd705c230 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.965578] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] d66c404d-8fd1-4fb7-a3b9-f21854c7e735/e7bd1d77-bfff-4684-9545-f6a1c69efe58-rescue.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.970225] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-14bd9583-8eb2-467c-9ce8-0636795a819e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.995571] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': task-1222569, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.997581] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 865.997581] env[62600]: value = "task-1222570" [ 865.997581] env[62600]: _type = "Task" [ 865.997581] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.009656] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222570, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.069758] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00308be1-99ec-435e-8814-ea6df12d87ab {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.077799] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f29ed35c-e292-416f-b2ee-f0af1cc82df7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.111720] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7904faea-3219-40e4-95c5-39caa7cda98d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.120685] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5254abe4-89b6-464f-905d-9f491c2eb42c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.137161] env[62600]: DEBUG nova.compute.provider_tree [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.466605] env[62600]: DEBUG oslo_vmware.api [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': task-1222569, 'name': PowerOnVM_Task, 'duration_secs': 0.640825} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.466915] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 866.467804] env[62600]: INFO nova.compute.manager [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Took 9.85 seconds to spawn the instance on the hypervisor. [ 866.467804] env[62600]: DEBUG nova.compute.manager [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 866.468709] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fea5ef1-d3b5-4b4a-8c1a-a0e3ec1b0e46 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.508591] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222570, 'name': ReconfigVM_Task, 'duration_secs': 0.297749} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.509084] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Reconfigured VM instance instance-00000042 to attach disk [datastore1] d66c404d-8fd1-4fb7-a3b9-f21854c7e735/e7bd1d77-bfff-4684-9545-f6a1c69efe58-rescue.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.512545] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d0c662c-4e1a-4160-89ed-e4280aebe329 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.541735] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-287eb52e-b239-4160-9815-0f116eb35db6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.558236] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 866.558236] env[62600]: value = "task-1222571" [ 866.558236] env[62600]: _type = "Task" [ 866.558236] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.568449] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222571, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.647718] env[62600]: DEBUG nova.scheduler.client.report [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 866.658164] env[62600]: DEBUG nova.network.neutron [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Successfully updated port: a0ecf848-49ae-489e-9c07-0acfa758d69f {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 866.991466] env[62600]: INFO nova.compute.manager [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Took 40.76 seconds to build instance. [ 867.068564] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222571, 'name': ReconfigVM_Task, 'duration_secs': 0.169914} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.069227] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 867.069227] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0b94590e-8bd1-49fa-a466-090aed86d993 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.076883] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 867.076883] env[62600]: value = "task-1222572" [ 867.076883] env[62600]: _type = "Task" [ 867.076883] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.084981] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222572, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.157010] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.949s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.167019] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "refresh_cache-bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.167019] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "refresh_cache-bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.167019] env[62600]: DEBUG nova.network.neutron [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 867.167019] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.244s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.167473] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.169144] env[62600]: DEBUG oslo_concurrency.lockutils [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.487s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.169493] env[62600]: DEBUG nova.objects.instance [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lazy-loading 'resources' on Instance uuid a358a3cb-deda-419a-aa3c-ce7aeb534240 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 867.192259] env[62600]: INFO nova.scheduler.client.report [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Deleted allocations for instance c2c20e5f-6c0e-4989-bc36-ff4a27d5c534 [ 867.203462] env[62600]: INFO nova.scheduler.client.report [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Deleted allocations for instance de4f9304-3357-4eaa-9c94-fe28bc554086 [ 867.396765] env[62600]: DEBUG oslo_concurrency.lockutils [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "d627a701-77cd-4a1d-9e52-0fc4adc93391" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.397089] env[62600]: DEBUG oslo_concurrency.lockutils [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "d627a701-77cd-4a1d-9e52-0fc4adc93391" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.496523] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2f34d4ca-30e9-4eae-92b5-20e871ff265c tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Lock "ddc80377-a14e-48d0-ad24-cc6daf7f6850" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.960s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.508452] env[62600]: DEBUG nova.compute.manager [req-1f2be715-4296-4250-a029-7bb82fb1f0b3 req-acf08853-b021-428a-ab12-c58f0a42b2ea service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Received event network-vif-plugged-a0ecf848-49ae-489e-9c07-0acfa758d69f {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.508452] env[62600]: DEBUG oslo_concurrency.lockutils [req-1f2be715-4296-4250-a029-7bb82fb1f0b3 req-acf08853-b021-428a-ab12-c58f0a42b2ea service nova] Acquiring lock "bd9b7351-01a2-429d-a860-d9ff6855eefc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.508567] env[62600]: DEBUG oslo_concurrency.lockutils [req-1f2be715-4296-4250-a029-7bb82fb1f0b3 req-acf08853-b021-428a-ab12-c58f0a42b2ea service nova] Lock "bd9b7351-01a2-429d-a860-d9ff6855eefc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.508738] env[62600]: DEBUG oslo_concurrency.lockutils [req-1f2be715-4296-4250-a029-7bb82fb1f0b3 req-acf08853-b021-428a-ab12-c58f0a42b2ea service nova] Lock "bd9b7351-01a2-429d-a860-d9ff6855eefc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.508976] env[62600]: DEBUG nova.compute.manager [req-1f2be715-4296-4250-a029-7bb82fb1f0b3 req-acf08853-b021-428a-ab12-c58f0a42b2ea service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] No waiting events found dispatching network-vif-plugged-a0ecf848-49ae-489e-9c07-0acfa758d69f {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 867.509122] env[62600]: WARNING nova.compute.manager [req-1f2be715-4296-4250-a029-7bb82fb1f0b3 req-acf08853-b021-428a-ab12-c58f0a42b2ea service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Received unexpected event network-vif-plugged-a0ecf848-49ae-489e-9c07-0acfa758d69f for instance with vm_state active and task_state None. [ 867.582058] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquiring lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.582058] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.592501] env[62600]: DEBUG oslo_vmware.api [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222572, 'name': PowerOnVM_Task, 'duration_secs': 0.380356} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.593037] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 867.596876] env[62600]: DEBUG nova.compute.manager [None req-030e8bb4-6138-4e0a-847d-89e7a30cbbef tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.598073] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f127cbb5-6e01-4ef8-a133-a65f417a42a2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.704555] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2b46d562-47c8-45bd-bfae-5efaa7fc509e tempest-ServerShowV257Test-1514396155 tempest-ServerShowV257Test-1514396155-project-member] Lock "c2c20e5f-6c0e-4989-bc36-ff4a27d5c534" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.062s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.724024] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ab9ec92-cc4b-4961-bb6e-fbe09fc95c49 tempest-MigrationsAdminTest-550072236 tempest-MigrationsAdminTest-550072236-project-member] Lock "de4f9304-3357-4eaa-9c94-fe28bc554086" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.969s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.904123] env[62600]: DEBUG nova.compute.manager [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 867.940949] env[62600]: WARNING nova.network.neutron [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] 8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb already exists in list: networks containing: ['8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb']. ignoring it [ 867.993344] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-211528b1-4a27-4d0e-99ff-069473a60e06 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.005561] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a67c1f1-52cc-4640-81ca-8eb700df634a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.042199] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea209cbb-b72d-4f71-b2bc-4aecf5fd496d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.051459] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3cbd149-aaa9-4f24-ba09-9975df8e557d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.067704] env[62600]: DEBUG nova.compute.provider_tree [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.082785] env[62600]: DEBUG nova.compute.utils [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 868.430588] env[62600]: DEBUG oslo_concurrency.lockutils [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.507331] env[62600]: DEBUG nova.network.neutron [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Updating instance_info_cache with network_info: [{"id": "75e67d5c-8153-4a74-b875-19d68778a85a", "address": "fa:16:3e:fa:9d:b3", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75e67d5c-81", "ovs_interfaceid": "75e67d5c-8153-4a74-b875-19d68778a85a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a0ecf848-49ae-489e-9c07-0acfa758d69f", "address": "fa:16:3e:df:28:f6", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0ecf848-49", "ovs_interfaceid": "a0ecf848-49ae-489e-9c07-0acfa758d69f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.572632] env[62600]: DEBUG nova.scheduler.client.report [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 868.588697] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.012728] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "refresh_cache-bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.013507] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.013694] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.014607] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d806530b-40af-489f-a87a-c5462f58b9f7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.041614] env[62600]: DEBUG nova.virt.hardware [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 869.042282] env[62600]: DEBUG nova.virt.hardware [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 869.042584] env[62600]: DEBUG nova.virt.hardware [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 869.043382] env[62600]: DEBUG nova.virt.hardware [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 869.043843] env[62600]: DEBUG nova.virt.hardware [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 869.045328] env[62600]: DEBUG nova.virt.hardware [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 869.045328] env[62600]: DEBUG nova.virt.hardware [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 869.045328] env[62600]: DEBUG nova.virt.hardware [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 869.045328] env[62600]: DEBUG nova.virt.hardware [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 869.045328] env[62600]: DEBUG nova.virt.hardware [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 869.045328] env[62600]: DEBUG nova.virt.hardware [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 869.052467] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Reconfiguring VM to attach interface {{(pid=62600) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 869.052976] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-811cdbd8-c532-46e0-8fe9-7333bfed690e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.074293] env[62600]: DEBUG oslo_vmware.api [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 869.074293] env[62600]: value = "task-1222573" [ 869.074293] env[62600]: _type = "Task" [ 869.074293] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.079038] env[62600]: DEBUG oslo_concurrency.lockutils [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.909s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.080486] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 16.724s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.080744] env[62600]: DEBUG nova.objects.instance [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62600) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 869.091256] env[62600]: DEBUG oslo_vmware.api [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222573, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.121797] env[62600]: INFO nova.scheduler.client.report [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Deleted allocations for instance a358a3cb-deda-419a-aa3c-ce7aeb534240 [ 869.592622] env[62600]: DEBUG oslo_vmware.api [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222573, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.631220] env[62600]: DEBUG oslo_concurrency.lockutils [None req-575a8558-2fb3-4803-ac6e-478f93c9a4a4 tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lock "a358a3cb-deda-419a-aa3c-ce7aeb534240" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.101s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.695251] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquiring lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.695575] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.695829] env[62600]: INFO nova.compute.manager [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Attaching volume 93edc5a4-1bb4-44c9-868a-43a307466fe0 to /dev/sdb [ 869.748704] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efcb29b4-4ae1-477c-8936-008bb7b64c85 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.759405] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac6a721c-e001-46a0-8157-2b2483edb584 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.777857] env[62600]: DEBUG nova.virt.block_device [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Updating existing volume attachment record: bcc925fd-1b4a-4562-bcba-af6e01041a22 {{(pid=62600) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 870.086898] env[62600]: DEBUG oslo_vmware.api [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222573, 'name': ReconfigVM_Task, 'duration_secs': 0.602941} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.087731] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.087964] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Reconfigured VM to attach interface {{(pid=62600) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 870.094595] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b8b902f-ba62-4d72-914f-1886e80575d6 tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.095625] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.193s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.100185] env[62600]: INFO nova.compute.claims [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 870.593870] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9deafdb8-20e1-4d11-ae42-f61a99bfa43a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "interface-bd9b7351-01a2-429d-a860-d9ff6855eefc-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.625s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.937024] env[62600]: DEBUG nova.compute.manager [req-cc5441fd-d4a1-4e93-9df4-14b24434ca17 req-f77955fc-7af7-4042-bdd5-af039f6eac76 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Received event network-changed-a0ecf848-49ae-489e-9c07-0acfa758d69f {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.937244] env[62600]: DEBUG nova.compute.manager [req-cc5441fd-d4a1-4e93-9df4-14b24434ca17 req-f77955fc-7af7-4042-bdd5-af039f6eac76 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Refreshing instance network info cache due to event network-changed-a0ecf848-49ae-489e-9c07-0acfa758d69f. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 870.937473] env[62600]: DEBUG oslo_concurrency.lockutils [req-cc5441fd-d4a1-4e93-9df4-14b24434ca17 req-f77955fc-7af7-4042-bdd5-af039f6eac76 service nova] Acquiring lock "refresh_cache-bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.937624] env[62600]: DEBUG oslo_concurrency.lockutils [req-cc5441fd-d4a1-4e93-9df4-14b24434ca17 req-f77955fc-7af7-4042-bdd5-af039f6eac76 service nova] Acquired lock "refresh_cache-bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.938192] env[62600]: DEBUG nova.network.neutron [req-cc5441fd-d4a1-4e93-9df4-14b24434ca17 req-f77955fc-7af7-4042-bdd5-af039f6eac76 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Refreshing network info cache for port a0ecf848-49ae-489e-9c07-0acfa758d69f {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 871.339552] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86887d85-2e36-4f7e-9ba9-8c00c58f49ed {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.347383] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb5d485e-a0e8-4a93-864d-39866ee20b29 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.389380] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4229de25-a0bb-417f-af18-b4931e6d5be2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.401212] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a86764-1d52-46ff-9f6d-67df5bdbf3cd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.417830] env[62600]: DEBUG nova.compute.provider_tree [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.469706] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "d2e43ba2-4ccf-4625-91ba-78a6f2632461" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.469997] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "d2e43ba2-4ccf-4625-91ba-78a6f2632461" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.504035] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "48bbbdaf-496d-4780-b467-6b393257535d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.504295] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "48bbbdaf-496d-4780-b467-6b393257535d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.678491] env[62600]: DEBUG nova.network.neutron [req-cc5441fd-d4a1-4e93-9df4-14b24434ca17 req-f77955fc-7af7-4042-bdd5-af039f6eac76 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Updated VIF entry in instance network info cache for port a0ecf848-49ae-489e-9c07-0acfa758d69f. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 871.678919] env[62600]: DEBUG nova.network.neutron [req-cc5441fd-d4a1-4e93-9df4-14b24434ca17 req-f77955fc-7af7-4042-bdd5-af039f6eac76 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Updating instance_info_cache with network_info: [{"id": "75e67d5c-8153-4a74-b875-19d68778a85a", "address": "fa:16:3e:fa:9d:b3", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75e67d5c-81", "ovs_interfaceid": "75e67d5c-8153-4a74-b875-19d68778a85a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a0ecf848-49ae-489e-9c07-0acfa758d69f", "address": "fa:16:3e:df:28:f6", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0ecf848-49", "ovs_interfaceid": "a0ecf848-49ae-489e-9c07-0acfa758d69f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.896448] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquiring lock "efff4d72-4673-4a8d-9a81-be9d700ff881" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.897062] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lock "efff4d72-4673-4a8d-9a81-be9d700ff881" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.897062] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquiring lock "efff4d72-4673-4a8d-9a81-be9d700ff881-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.897330] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lock "efff4d72-4673-4a8d-9a81-be9d700ff881-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.897714] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lock "efff4d72-4673-4a8d-9a81-be9d700ff881-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.901542] env[62600]: INFO nova.compute.manager [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Terminating instance [ 871.905219] env[62600]: DEBUG nova.compute.manager [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 871.905219] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 871.905401] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ef4cfee-1a98-4bb3-b92e-ef7553440814 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.914784] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 871.915168] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-46475409-8e28-4b7d-958a-4c6b9a18e90f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.920341] env[62600]: DEBUG nova.scheduler.client.report [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 871.925978] env[62600]: DEBUG oslo_vmware.api [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 871.925978] env[62600]: value = "task-1222577" [ 871.925978] env[62600]: _type = "Task" [ 871.925978] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.938629] env[62600]: DEBUG oslo_vmware.api [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222577, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.973424] env[62600]: DEBUG nova.compute.manager [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 872.009055] env[62600]: DEBUG nova.compute.manager [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 872.183744] env[62600]: DEBUG oslo_concurrency.lockutils [req-cc5441fd-d4a1-4e93-9df4-14b24434ca17 req-f77955fc-7af7-4042-bdd5-af039f6eac76 service nova] Releasing lock "refresh_cache-bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.428499] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.429111] env[62600]: DEBUG nova.compute.manager [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 872.434588] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.329s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.435035] env[62600]: DEBUG nova.objects.instance [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lazy-loading 'resources' on Instance uuid af7036df-b9f0-4ce6-962a-1edd7c1ea211 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 872.451737] env[62600]: DEBUG oslo_vmware.api [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222577, 'name': PowerOffVM_Task, 'duration_secs': 0.481517} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.452098] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 872.452284] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 872.452836] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0d3604c0-e922-4817-8c36-3d34c4811b7c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.507032] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.540460] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.551321] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 872.551572] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 872.551776] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Deleting the datastore file [datastore2] efff4d72-4673-4a8d-9a81-be9d700ff881 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 872.552131] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4147aefc-71f9-424c-9b6b-889fcdbba6ff {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.563408] env[62600]: DEBUG oslo_vmware.api [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for the task: (returnval){ [ 872.563408] env[62600]: value = "task-1222580" [ 872.563408] env[62600]: _type = "Task" [ 872.563408] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.574449] env[62600]: DEBUG oslo_vmware.api [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222580, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.714809] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 872.714809] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 872.873676] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "interface-bd9b7351-01a2-429d-a860-d9ff6855eefc-a0ecf848-49ae-489e-9c07-0acfa758d69f" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.873676] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "interface-bd9b7351-01a2-429d-a860-d9ff6855eefc-a0ecf848-49ae-489e-9c07-0acfa758d69f" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.936525] env[62600]: DEBUG nova.compute.utils [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 872.938163] env[62600]: DEBUG nova.compute.manager [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 872.940923] env[62600]: DEBUG nova.network.neutron [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 872.944235] env[62600]: DEBUG nova.objects.instance [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lazy-loading 'numa_topology' on Instance uuid af7036df-b9f0-4ce6-962a-1edd7c1ea211 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 873.012217] env[62600]: DEBUG nova.policy [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3e59998217a4b18a6f3f01142a5e440', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f91091f83ee4a2091507ca994e3d52f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 873.081075] env[62600]: DEBUG oslo_vmware.api [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Task: {'id': task-1222580, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.28299} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.081981] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 873.081981] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 873.082395] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 873.082749] env[62600]: INFO nova.compute.manager [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Took 1.18 seconds to destroy the instance on the hypervisor. [ 873.083210] env[62600]: DEBUG oslo.service.loopingcall [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.083557] env[62600]: DEBUG nova.compute.manager [-] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 873.083761] env[62600]: DEBUG nova.network.neutron [-] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 873.222110] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 873.222965] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Starting heal instance info cache {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 873.222965] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Rebuilding the list of instances to heal {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 873.376438] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.376632] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.377604] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eba2852-c83b-491c-a017-c8e435055b46 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.408031] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40eed80f-cc3f-4c9a-9ca4-3a33f55c4333 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.440694] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Reconfiguring VM to detach interface {{(pid=62600) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 873.441634] env[62600]: DEBUG nova.compute.manager [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 873.444145] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ffe831a3-c0aa-42b1-b143-3561750b8a17 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.463683] env[62600]: DEBUG nova.network.neutron [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Successfully created port: 30cc50bd-a129-4992-8ced-66359d34fcc1 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 873.467034] env[62600]: DEBUG nova.objects.base [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62600) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 873.476970] env[62600]: DEBUG oslo_vmware.api [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 873.476970] env[62600]: value = "task-1222581" [ 873.476970] env[62600]: _type = "Task" [ 873.476970] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.486525] env[62600]: DEBUG oslo_vmware.api [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222581, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.614187] env[62600]: DEBUG nova.compute.manager [req-b613a281-353d-4b2c-88e3-27b639dc7d65 req-3d21f405-134e-47c1-b109-9b36c44d82da service nova] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Received event network-vif-deleted-c5d84e4b-f903-4d3b-ad0f-4a596b324f41 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.614187] env[62600]: INFO nova.compute.manager [req-b613a281-353d-4b2c-88e3-27b639dc7d65 req-3d21f405-134e-47c1-b109-9b36c44d82da service nova] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Neutron deleted interface c5d84e4b-f903-4d3b-ad0f-4a596b324f41; detaching it from the instance and deleting it from the info cache [ 873.614187] env[62600]: DEBUG nova.network.neutron [req-b613a281-353d-4b2c-88e3-27b639dc7d65 req-3d21f405-134e-47c1-b109-9b36c44d82da service nova] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.729864] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Skipping network cache update for instance because it is being deleted. {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 873.730130] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Skipping network cache update for instance because it is being deleted. {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 873.730130] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Skipping network cache update for instance because it is Building. {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 873.764506] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa6439a-4292-4b7f-ac81-54080e492275 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.775340] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63720f25-d9f1-44d0-a3e5-624db9f68235 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.814298] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "refresh_cache-afe838cc-e086-4986-87ec-4e1266bcaf60" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.814502] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquired lock "refresh_cache-afe838cc-e086-4986-87ec-4e1266bcaf60" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.814636] env[62600]: DEBUG nova.network.neutron [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Forcefully refreshing network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 873.815214] env[62600]: DEBUG nova.objects.instance [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lazy-loading 'info_cache' on Instance uuid afe838cc-e086-4986-87ec-4e1266bcaf60 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 873.818355] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf22f1d-f793-4b01-8438-3ff3851480c1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.827996] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6417346-cf54-4ba1-b774-b447526f508b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.844617] env[62600]: DEBUG nova.compute.provider_tree [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.859833] env[62600]: INFO nova.compute.manager [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Rescuing [ 873.860041] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "refresh_cache-b967fb11-e70e-4e17-b769-38da581bd83b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.860191] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquired lock "refresh_cache-b967fb11-e70e-4e17-b769-38da581bd83b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.860371] env[62600]: DEBUG nova.network.neutron [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 873.976028] env[62600]: DEBUG nova.network.neutron [-] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.988573] env[62600]: DEBUG oslo_vmware.api [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.079924] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Acquiring lock "4eaa22b9-5834-4425-82a9-d16e316f3a52" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.080206] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Lock "4eaa22b9-5834-4425-82a9-d16e316f3a52" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.118892] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c02c620b-bfa7-49a4-8624-6329c2464871 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.130443] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a7d1e12-3fa8-48e8-a311-24c0b7c5dfc6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.160104] env[62600]: DEBUG nova.compute.manager [req-b613a281-353d-4b2c-88e3-27b639dc7d65 req-3d21f405-134e-47c1-b109-9b36c44d82da service nova] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Detach interface failed, port_id=c5d84e4b-f903-4d3b-ad0f-4a596b324f41, reason: Instance efff4d72-4673-4a8d-9a81-be9d700ff881 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 874.333016] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Volume attach. Driver type: vmdk {{(pid=62600) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 874.333308] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264299', 'volume_id': '93edc5a4-1bb4-44c9-868a-43a307466fe0', 'name': 'volume-93edc5a4-1bb4-44c9-868a-43a307466fe0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dde6df10-618a-40a8-b33f-efc0ca3a9287', 'attached_at': '', 'detached_at': '', 'volume_id': '93edc5a4-1bb4-44c9-868a-43a307466fe0', 'serial': '93edc5a4-1bb4-44c9-868a-43a307466fe0'} {{(pid=62600) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 874.334421] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-435ad14f-eda3-4d36-be7b-4fcaf27e9020 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.351494] env[62600]: DEBUG nova.scheduler.client.report [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 874.355505] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-186add7f-7397-44d2-b6bd-a333676377eb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.383308] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] volume-93edc5a4-1bb4-44c9-868a-43a307466fe0/volume-93edc5a4-1bb4-44c9-868a-43a307466fe0.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 874.383868] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0d76115-b902-4fd0-973a-b9de3d276cdd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.405154] env[62600]: DEBUG oslo_vmware.api [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for the task: (returnval){ [ 874.405154] env[62600]: value = "task-1222582" [ 874.405154] env[62600]: _type = "Task" [ 874.405154] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.415107] env[62600]: DEBUG oslo_vmware.api [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222582, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.417600] env[62600]: DEBUG oslo_concurrency.lockutils [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Acquiring lock "ddc80377-a14e-48d0-ad24-cc6daf7f6850" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.417826] env[62600]: DEBUG oslo_concurrency.lockutils [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Lock "ddc80377-a14e-48d0-ad24-cc6daf7f6850" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.418130] env[62600]: DEBUG oslo_concurrency.lockutils [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Acquiring lock "ddc80377-a14e-48d0-ad24-cc6daf7f6850-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.418339] env[62600]: DEBUG oslo_concurrency.lockutils [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Lock "ddc80377-a14e-48d0-ad24-cc6daf7f6850-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.418515] env[62600]: DEBUG oslo_concurrency.lockutils [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Lock "ddc80377-a14e-48d0-ad24-cc6daf7f6850-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.420635] env[62600]: INFO nova.compute.manager [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Terminating instance [ 874.422591] env[62600]: DEBUG nova.compute.manager [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 874.422800] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 874.424307] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8894bb4-045d-401d-8b99-770a0dbe62e3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.431571] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 874.432093] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-baccf587-d568-4649-8ac3-df632f5ffd5b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.439198] env[62600]: DEBUG oslo_vmware.api [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Waiting for the task: (returnval){ [ 874.439198] env[62600]: value = "task-1222583" [ 874.439198] env[62600]: _type = "Task" [ 874.439198] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.448033] env[62600]: DEBUG oslo_vmware.api [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': task-1222583, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.476623] env[62600]: DEBUG nova.compute.manager [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 874.479929] env[62600]: INFO nova.compute.manager [-] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Took 1.40 seconds to deallocate network for instance. [ 874.496329] env[62600]: DEBUG oslo_vmware.api [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.520402] env[62600]: DEBUG nova.virt.hardware [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 874.520402] env[62600]: DEBUG nova.virt.hardware [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 874.520402] env[62600]: DEBUG nova.virt.hardware [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.520402] env[62600]: DEBUG nova.virt.hardware [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 874.520402] env[62600]: DEBUG nova.virt.hardware [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.520402] env[62600]: DEBUG nova.virt.hardware [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 874.520853] env[62600]: DEBUG nova.virt.hardware [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 874.520853] env[62600]: DEBUG nova.virt.hardware [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 874.520978] env[62600]: DEBUG nova.virt.hardware [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 874.521068] env[62600]: DEBUG nova.virt.hardware [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 874.521254] env[62600]: DEBUG nova.virt.hardware [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 874.522160] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab3e4ac-d093-4262-bb31-ea7abc3e1306 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.533031] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e66ed3-d1f1-465b-b8e3-1ed839485bfb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.582679] env[62600]: DEBUG nova.compute.manager [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 874.795703] env[62600]: DEBUG nova.network.neutron [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Updating instance_info_cache with network_info: [{"id": "d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049", "address": "fa:16:3e:4f:fd:77", "network": {"id": "036dc114-6ea0-46cb-a958-4a6137e7e85f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-728627542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e1772c9a6c4671ab89df2e8b40cae1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6e6a2fb-7b", "ovs_interfaceid": "d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.855843] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.423s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.858512] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.222s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.861633] env[62600]: DEBUG nova.objects.instance [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Lazy-loading 'resources' on Instance uuid ee7175b9-dbe6-4f90-bd2b-8829194dc6c3 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.917722] env[62600]: DEBUG oslo_vmware.api [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222582, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.951489] env[62600]: DEBUG oslo_vmware.api [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': task-1222583, 'name': PowerOffVM_Task, 'duration_secs': 0.184626} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.951668] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 874.951911] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 874.952243] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-324a2e85-9ddf-41c8-a9a3-3ea94aa75700 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.989858] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.995144] env[62600]: DEBUG oslo_vmware.api [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.042808] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 875.043039] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 875.043282] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Deleting the datastore file [datastore1] ddc80377-a14e-48d0-ad24-cc6daf7f6850 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 875.043563] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-862fd5e6-1bb3-475c-9f4a-eb13a4fd351b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.052069] env[62600]: DEBUG oslo_vmware.api [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Waiting for the task: (returnval){ [ 875.052069] env[62600]: value = "task-1222585" [ 875.052069] env[62600]: _type = "Task" [ 875.052069] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.060822] env[62600]: DEBUG oslo_vmware.api [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': task-1222585, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.111022] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.177095] env[62600]: DEBUG nova.network.neutron [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Successfully updated port: 30cc50bd-a129-4992-8ced-66359d34fcc1 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 875.301231] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Releasing lock "refresh_cache-b967fb11-e70e-4e17-b769-38da581bd83b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.368017] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7afbd939-85e9-4e72-a07c-4a9e31c4748f tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "af7036df-b9f0-4ce6-962a-1edd7c1ea211" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 40.515s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.368909] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0a69bddc-7b26-47dd-b61b-af438cda10d1 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "af7036df-b9f0-4ce6-962a-1edd7c1ea211" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 18.509s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.369192] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0a69bddc-7b26-47dd-b61b-af438cda10d1 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "af7036df-b9f0-4ce6-962a-1edd7c1ea211-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.369470] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0a69bddc-7b26-47dd-b61b-af438cda10d1 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "af7036df-b9f0-4ce6-962a-1edd7c1ea211-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.369577] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0a69bddc-7b26-47dd-b61b-af438cda10d1 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "af7036df-b9f0-4ce6-962a-1edd7c1ea211-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.378917] env[62600]: INFO nova.compute.manager [None req-0a69bddc-7b26-47dd-b61b-af438cda10d1 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Terminating instance [ 875.383798] env[62600]: DEBUG nova.compute.manager [None req-0a69bddc-7b26-47dd-b61b-af438cda10d1 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 875.383798] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-0a69bddc-7b26-47dd-b61b-af438cda10d1 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 875.384885] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d4e37dd4-472a-4a60-aeca-4527b2d03bab {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.397113] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a08e32ef-13a7-4a88-af9a-744fa9180181 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.430213] env[62600]: DEBUG oslo_vmware.api [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222582, 'name': ReconfigVM_Task, 'duration_secs': 0.58781} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.440229] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Reconfigured VM instance instance-0000003b to attach disk [datastore2] volume-93edc5a4-1bb4-44c9-868a-43a307466fe0/volume-93edc5a4-1bb4-44c9-868a-43a307466fe0.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 875.445266] env[62600]: WARNING nova.virt.vmwareapi.vmops [None req-0a69bddc-7b26-47dd-b61b-af438cda10d1 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance af7036df-b9f0-4ce6-962a-1edd7c1ea211 could not be found. [ 875.445469] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-0a69bddc-7b26-47dd-b61b-af438cda10d1 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.445657] env[62600]: INFO nova.compute.manager [None req-0a69bddc-7b26-47dd-b61b-af438cda10d1 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Took 0.06 seconds to destroy the instance on the hypervisor. [ 875.445893] env[62600]: DEBUG oslo.service.loopingcall [None req-0a69bddc-7b26-47dd-b61b-af438cda10d1 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.452427] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f2fa131-44f1-4144-967e-923c13046226 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.460686] env[62600]: DEBUG nova.compute.manager [-] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 875.460803] env[62600]: DEBUG nova.network.neutron [-] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 875.476813] env[62600]: DEBUG oslo_vmware.api [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for the task: (returnval){ [ 875.476813] env[62600]: value = "task-1222586" [ 875.476813] env[62600]: _type = "Task" [ 875.476813] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.489177] env[62600]: DEBUG oslo_vmware.api [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222586, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.498809] env[62600]: DEBUG oslo_vmware.api [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.567397] env[62600]: DEBUG oslo_vmware.api [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Task: {'id': task-1222585, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177004} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.570157] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.571149] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 875.571149] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.571149] env[62600]: INFO nova.compute.manager [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Took 1.15 seconds to destroy the instance on the hypervisor. [ 875.571149] env[62600]: DEBUG oslo.service.loopingcall [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.572353] env[62600]: DEBUG nova.compute.manager [-] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 875.572353] env[62600]: DEBUG nova.network.neutron [-] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 875.655476] env[62600]: DEBUG nova.compute.manager [req-ba934247-4577-459b-925f-c59c0d99691b req-9c14bf8f-be95-4ceb-8cb7-256e4252c4e3 service nova] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Received event network-vif-plugged-30cc50bd-a129-4992-8ced-66359d34fcc1 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.655706] env[62600]: DEBUG oslo_concurrency.lockutils [req-ba934247-4577-459b-925f-c59c0d99691b req-9c14bf8f-be95-4ceb-8cb7-256e4252c4e3 service nova] Acquiring lock "1b856aff-66be-4fa5-b8b7-124ac24fb4d1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.655926] env[62600]: DEBUG oslo_concurrency.lockutils [req-ba934247-4577-459b-925f-c59c0d99691b req-9c14bf8f-be95-4ceb-8cb7-256e4252c4e3 service nova] Lock "1b856aff-66be-4fa5-b8b7-124ac24fb4d1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.656114] env[62600]: DEBUG oslo_concurrency.lockutils [req-ba934247-4577-459b-925f-c59c0d99691b req-9c14bf8f-be95-4ceb-8cb7-256e4252c4e3 service nova] Lock "1b856aff-66be-4fa5-b8b7-124ac24fb4d1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.656289] env[62600]: DEBUG nova.compute.manager [req-ba934247-4577-459b-925f-c59c0d99691b req-9c14bf8f-be95-4ceb-8cb7-256e4252c4e3 service nova] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] No waiting events found dispatching network-vif-plugged-30cc50bd-a129-4992-8ced-66359d34fcc1 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 875.656458] env[62600]: WARNING nova.compute.manager [req-ba934247-4577-459b-925f-c59c0d99691b req-9c14bf8f-be95-4ceb-8cb7-256e4252c4e3 service nova] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Received unexpected event network-vif-plugged-30cc50bd-a129-4992-8ced-66359d34fcc1 for instance with vm_state building and task_state spawning. [ 875.656619] env[62600]: DEBUG nova.compute.manager [req-ba934247-4577-459b-925f-c59c0d99691b req-9c14bf8f-be95-4ceb-8cb7-256e4252c4e3 service nova] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Received event network-changed-30cc50bd-a129-4992-8ced-66359d34fcc1 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.656803] env[62600]: DEBUG nova.compute.manager [req-ba934247-4577-459b-925f-c59c0d99691b req-9c14bf8f-be95-4ceb-8cb7-256e4252c4e3 service nova] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Refreshing instance network info cache due to event network-changed-30cc50bd-a129-4992-8ced-66359d34fcc1. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 875.656949] env[62600]: DEBUG oslo_concurrency.lockutils [req-ba934247-4577-459b-925f-c59c0d99691b req-9c14bf8f-be95-4ceb-8cb7-256e4252c4e3 service nova] Acquiring lock "refresh_cache-1b856aff-66be-4fa5-b8b7-124ac24fb4d1" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.657106] env[62600]: DEBUG oslo_concurrency.lockutils [req-ba934247-4577-459b-925f-c59c0d99691b req-9c14bf8f-be95-4ceb-8cb7-256e4252c4e3 service nova] Acquired lock "refresh_cache-1b856aff-66be-4fa5-b8b7-124ac24fb4d1" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.657267] env[62600]: DEBUG nova.network.neutron [req-ba934247-4577-459b-925f-c59c0d99691b req-9c14bf8f-be95-4ceb-8cb7-256e4252c4e3 service nova] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Refreshing network info cache for port 30cc50bd-a129-4992-8ced-66359d34fcc1 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 875.684643] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "refresh_cache-1b856aff-66be-4fa5-b8b7-124ac24fb4d1" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.706185] env[62600]: DEBUG nova.network.neutron [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Updating instance_info_cache with network_info: [{"id": "b4b6687a-b518-40c1-99a5-7d1ffe49e1f2", "address": "fa:16:3e:2a:79:1c", "network": {"id": "102eb76e-a1ff-4b27-b285-516053dcad35", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-543226554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "888577fb598f45c881757ce3ee52bfa3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4b6687a-b5", "ovs_interfaceid": "b4b6687a-b518-40c1-99a5-7d1ffe49e1f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.724268] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6277416-f185-4f1f-9103-971b2578f742 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.733814] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f14a56-1d95-45de-8630-6accc7acbab1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.764506] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9635663f-424e-4d1f-97db-25c97832a703 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.773607] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd0d473-6fbd-4113-b773-b0ca7bbf23a4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.790217] env[62600]: DEBUG nova.compute.provider_tree [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.844368] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 875.844658] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-867a33bd-ead3-4c71-9d19-9e1bba87da9a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.855233] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 875.855233] env[62600]: value = "task-1222587" [ 875.855233] env[62600]: _type = "Task" [ 875.855233] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.865822] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222587, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.993029] env[62600]: DEBUG oslo_vmware.api [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222586, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.996378] env[62600]: DEBUG oslo_vmware.api [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.213555] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Releasing lock "refresh_cache-afe838cc-e086-4986-87ec-4e1266bcaf60" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.213555] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Updated the network info_cache for instance {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 876.213555] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 876.213555] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 876.213555] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 876.213555] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 876.213555] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 876.213555] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 876.213555] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62600) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 876.213555] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 876.221757] env[62600]: DEBUG nova.network.neutron [req-ba934247-4577-459b-925f-c59c0d99691b req-9c14bf8f-be95-4ceb-8cb7-256e4252c4e3 service nova] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 876.295361] env[62600]: DEBUG nova.scheduler.client.report [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 876.311397] env[62600]: DEBUG nova.network.neutron [req-ba934247-4577-459b-925f-c59c0d99691b req-9c14bf8f-be95-4ceb-8cb7-256e4252c4e3 service nova] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.344560] env[62600]: DEBUG nova.network.neutron [-] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.369542] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222587, 'name': PowerOffVM_Task, 'duration_secs': 0.319737} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.369828] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 876.370685] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290683c7-c737-4c61-936e-221c72f81a2e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.400854] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1254ccb8-2bc4-4063-8eac-ec6e70dcb833 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.444102] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 876.444788] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d8982eaa-e0ec-467c-82b1-8edb2bc4c26e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.454073] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 876.454073] env[62600]: value = "task-1222588" [ 876.454073] env[62600]: _type = "Task" [ 876.454073] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.465126] env[62600]: DEBUG nova.compute.manager [req-feda379e-93a5-4bc9-b246-99dcfe980cab req-d5b50097-0137-4b36-83da-b6adfce53522 service nova] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Received event network-vif-deleted-c493ab77-1118-42b6-8009-41c5eced4543 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 876.465299] env[62600]: INFO nova.compute.manager [req-feda379e-93a5-4bc9-b246-99dcfe980cab req-d5b50097-0137-4b36-83da-b6adfce53522 service nova] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Neutron deleted interface c493ab77-1118-42b6-8009-41c5eced4543; detaching it from the instance and deleting it from the info cache [ 876.465654] env[62600]: DEBUG nova.network.neutron [req-feda379e-93a5-4bc9-b246-99dcfe980cab req-d5b50097-0137-4b36-83da-b6adfce53522 service nova] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.475974] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] VM already powered off {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 876.475974] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.475974] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.475974] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.475974] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.476309] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c00fdc54-6a72-4c9b-b5af-a1d53699d407 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.496116] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.496301] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 876.500588] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-462ecc67-2fc8-4cad-b4ea-2cbf51b10d76 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.503679] env[62600]: DEBUG oslo_vmware.api [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222586, 'name': ReconfigVM_Task, 'duration_secs': 0.598575} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.507427] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264299', 'volume_id': '93edc5a4-1bb4-44c9-868a-43a307466fe0', 'name': 'volume-93edc5a4-1bb4-44c9-868a-43a307466fe0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dde6df10-618a-40a8-b33f-efc0ca3a9287', 'attached_at': '', 'detached_at': '', 'volume_id': '93edc5a4-1bb4-44c9-868a-43a307466fe0', 'serial': '93edc5a4-1bb4-44c9-868a-43a307466fe0'} {{(pid=62600) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 876.509300] env[62600]: DEBUG oslo_vmware.api [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.513838] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 876.513838] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52973af9-eda9-feea-ac01-b1b9d017f605" [ 876.513838] env[62600]: _type = "Task" [ 876.513838] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.522266] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52973af9-eda9-feea-ac01-b1b9d017f605, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.717784] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.801503] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.942s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.804668] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.092s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.806321] env[62600]: INFO nova.compute.claims [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 876.818771] env[62600]: DEBUG oslo_concurrency.lockutils [req-ba934247-4577-459b-925f-c59c0d99691b req-9c14bf8f-be95-4ceb-8cb7-256e4252c4e3 service nova] Releasing lock "refresh_cache-1b856aff-66be-4fa5-b8b7-124ac24fb4d1" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.819177] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "refresh_cache-1b856aff-66be-4fa5-b8b7-124ac24fb4d1" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.819359] env[62600]: DEBUG nova.network.neutron [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 876.827343] env[62600]: DEBUG nova.network.neutron [-] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.831017] env[62600]: INFO nova.scheduler.client.report [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Deleted allocations for instance ee7175b9-dbe6-4f90-bd2b-8829194dc6c3 [ 876.847889] env[62600]: INFO nova.compute.manager [-] [instance: af7036df-b9f0-4ce6-962a-1edd7c1ea211] Took 1.39 seconds to deallocate network for instance. [ 876.975379] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4cb173fb-4d33-4fbf-9614-9e6e6f37b254 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.986376] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63f87e34-20fa-440b-848b-2e527c856667 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.006063] env[62600]: DEBUG oslo_vmware.api [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.020030] env[62600]: DEBUG nova.compute.manager [req-feda379e-93a5-4bc9-b246-99dcfe980cab req-d5b50097-0137-4b36-83da-b6adfce53522 service nova] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Detach interface failed, port_id=c493ab77-1118-42b6-8009-41c5eced4543, reason: Instance ddc80377-a14e-48d0-ad24-cc6daf7f6850 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 877.029982] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52973af9-eda9-feea-ac01-b1b9d017f605, 'name': SearchDatastore_Task, 'duration_secs': 0.013312} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.030783] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f26a4a73-f567-475c-901c-c56f9ab33e1a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.036530] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 877.036530] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521d670c-2275-af59-0b17-28779dfee5ca" [ 877.036530] env[62600]: _type = "Task" [ 877.036530] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.044915] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521d670c-2275-af59-0b17-28779dfee5ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.329241] env[62600]: INFO nova.compute.manager [-] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Took 1.76 seconds to deallocate network for instance. [ 877.343622] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc49e8b3-5c6e-47d7-9c21-6874c04373bf tempest-ServerShowV254Test-2067756709 tempest-ServerShowV254Test-2067756709-project-member] Lock "ee7175b9-dbe6-4f90-bd2b-8829194dc6c3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.686s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.381724] env[62600]: DEBUG nova.network.neutron [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 877.508252] env[62600]: DEBUG oslo_vmware.api [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.547284] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521d670c-2275-af59-0b17-28779dfee5ca, 'name': SearchDatastore_Task, 'duration_secs': 0.010415} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.547576] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.547838] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] b967fb11-e70e-4e17-b769-38da581bd83b/e7bd1d77-bfff-4684-9545-f6a1c69efe58-rescue.vmdk. {{(pid=62600) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 877.548138] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-656e8354-437f-45a6-b482-e5b130fc586f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.553416] env[62600]: DEBUG nova.objects.instance [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lazy-loading 'flavor' on Instance uuid dde6df10-618a-40a8-b33f-efc0ca3a9287 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.559983] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 877.559983] env[62600]: value = "task-1222589" [ 877.559983] env[62600]: _type = "Task" [ 877.559983] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.571755] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222589, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.673365] env[62600]: DEBUG nova.network.neutron [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Updating instance_info_cache with network_info: [{"id": "30cc50bd-a129-4992-8ced-66359d34fcc1", "address": "fa:16:3e:83:f4:a4", "network": {"id": "1d2052cf-5be1-48e7-887a-18f367b7823f", "bridge": "br-int", "label": "tempest-ServersTestJSON-106099679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f91091f83ee4a2091507ca994e3d52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30cc50bd-a1", "ovs_interfaceid": "30cc50bd-a129-4992-8ced-66359d34fcc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.846228] env[62600]: DEBUG oslo_concurrency.lockutils [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.890533] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0a69bddc-7b26-47dd-b61b-af438cda10d1 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "af7036df-b9f0-4ce6-962a-1edd7c1ea211" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.521s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.012592] env[62600]: DEBUG oslo_vmware.api [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.059701] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b961459d-e89d-4b92-8228-f54ecf45edab tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.364s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.077743] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222589, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464924} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.078215] env[62600]: INFO nova.virt.vmwareapi.ds_util [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] b967fb11-e70e-4e17-b769-38da581bd83b/e7bd1d77-bfff-4684-9545-f6a1c69efe58-rescue.vmdk. [ 878.079269] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da07747c-5c26-4c62-aa31-9051380e3798 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.114016] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] b967fb11-e70e-4e17-b769-38da581bd83b/e7bd1d77-bfff-4684-9545-f6a1c69efe58-rescue.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 878.115830] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ceb6a544-840f-4916-9b5d-1dd1da9ef881 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.139223] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 878.139223] env[62600]: value = "task-1222590" [ 878.139223] env[62600]: _type = "Task" [ 878.139223] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.150220] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222590, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.182117] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "refresh_cache-1b856aff-66be-4fa5-b8b7-124ac24fb4d1" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.182117] env[62600]: DEBUG nova.compute.manager [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Instance network_info: |[{"id": "30cc50bd-a129-4992-8ced-66359d34fcc1", "address": "fa:16:3e:83:f4:a4", "network": {"id": "1d2052cf-5be1-48e7-887a-18f367b7823f", "bridge": "br-int", "label": "tempest-ServersTestJSON-106099679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f91091f83ee4a2091507ca994e3d52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30cc50bd-a1", "ovs_interfaceid": "30cc50bd-a129-4992-8ced-66359d34fcc1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 878.184508] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:f4:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '028bae2d-fe6c-4207-b4a3-3fab45fbf1d6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '30cc50bd-a129-4992-8ced-66359d34fcc1', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 878.195288] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Creating folder: Project (8f91091f83ee4a2091507ca994e3d52f). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 878.195419] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5aad7e01-88fd-49cf-adf6-01910873517c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.210941] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Created folder: Project (8f91091f83ee4a2091507ca994e3d52f) in parent group-v264198. [ 878.211467] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Creating folder: Instances. Parent ref: group-v264300. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 878.212330] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-454f5e31-09bc-49fb-84d9-c44719826e4c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.223271] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f647873c-174a-4870-a04b-b3b488260217 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.227734] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Created folder: Instances in parent group-v264300. [ 878.229102] env[62600]: DEBUG oslo.service.loopingcall [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.229102] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 878.229102] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8ffed277-652e-402a-bab8-c1bad61d9dee {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.248447] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ddbf0cd-d435-4de4-a7c6-1837f8acae60 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.254149] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 878.254149] env[62600]: value = "task-1222593" [ 878.254149] env[62600]: _type = "Task" [ 878.254149] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.285707] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e90a917f-825f-4e99-8965-1cd0158187b7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.292017] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222593, 'name': CreateVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.299239] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9adb8d9c-4aa9-486f-a5c2-050769f29cc2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.314166] env[62600]: DEBUG nova.compute.provider_tree [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.508218] env[62600]: DEBUG oslo_vmware.api [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.649531] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222590, 'name': ReconfigVM_Task, 'duration_secs': 0.410832} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.650148] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Reconfigured VM instance instance-00000041 to attach disk [datastore2] b967fb11-e70e-4e17-b769-38da581bd83b/e7bd1d77-bfff-4684-9545-f6a1c69efe58-rescue.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 878.651453] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab7eb0c-897e-4605-a5e1-903e6ad8f3a4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.680160] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ec3ce97-b706-434c-a452-c11fba15e75e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.701338] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 878.701338] env[62600]: value = "task-1222594" [ 878.701338] env[62600]: _type = "Task" [ 878.701338] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.710749] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222594, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.764898] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222593, 'name': CreateVM_Task, 'duration_secs': 0.448853} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.764898] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 878.765919] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.765919] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.765919] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 878.766152] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc0757ed-4dac-45a7-98cb-55b7f5610f61 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.774029] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 878.774029] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5270d344-400e-81cd-b328-d0882454a42f" [ 878.774029] env[62600]: _type = "Task" [ 878.774029] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.788233] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5270d344-400e-81cd-b328-d0882454a42f, 'name': SearchDatastore_Task, 'duration_secs': 0.01093} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.788558] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.788873] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 878.789181] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.789379] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.789655] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 878.789951] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-01bfd919-3715-4039-8551-633e2c72f4e4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.801434] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 878.801699] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 878.802551] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7dd384f8-d355-4175-bdcd-6a84c401f7a6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.808954] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 878.808954] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52516fe6-d563-9c51-2eaa-777c3cfaa59f" [ 878.808954] env[62600]: _type = "Task" [ 878.808954] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.820493] env[62600]: DEBUG nova.scheduler.client.report [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 878.825162] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52516fe6-d563-9c51-2eaa-777c3cfaa59f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.911096] env[62600]: DEBUG oslo_concurrency.lockutils [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "043ac602-6c4a-4275-b019-5d717579e7c0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.911374] env[62600]: DEBUG oslo_concurrency.lockutils [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "043ac602-6c4a-4275-b019-5d717579e7c0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.008268] env[62600]: DEBUG oslo_vmware.api [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222581, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.211764] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222594, 'name': ReconfigVM_Task, 'duration_secs': 0.178647} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.212064] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 879.212337] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ebb6ddbe-862d-46ac-b502-5974317af2c3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.221879] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 879.221879] env[62600]: value = "task-1222595" [ 879.221879] env[62600]: _type = "Task" [ 879.221879] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.230554] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222595, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.286996] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquiring lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.286996] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.321013] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52516fe6-d563-9c51-2eaa-777c3cfaa59f, 'name': SearchDatastore_Task, 'duration_secs': 0.013156} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.322239] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c647ed1c-8f49-453a-bc87-51136f996913 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.325553] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.326140] env[62600]: DEBUG nova.compute.manager [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 879.330829] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.330s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.331983] env[62600]: INFO nova.compute.claims [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 879.335483] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 879.335483] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52daeda0-32bc-3f09-9558-e6013b392708" [ 879.335483] env[62600]: _type = "Task" [ 879.335483] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.349012] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52daeda0-32bc-3f09-9558-e6013b392708, 'name': SearchDatastore_Task, 'duration_secs': 0.010196} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.349012] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.349012] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 1b856aff-66be-4fa5-b8b7-124ac24fb4d1/1b856aff-66be-4fa5-b8b7-124ac24fb4d1.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 879.349012] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8a99cf4c-7b66-4cb1-beeb-a4c7f3022484 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.357150] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 879.357150] env[62600]: value = "task-1222596" [ 879.357150] env[62600]: _type = "Task" [ 879.357150] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.365484] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222596, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.415468] env[62600]: DEBUG nova.compute.manager [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 879.509533] env[62600]: DEBUG oslo_vmware.api [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222581, 'name': ReconfigVM_Task, 'duration_secs': 5.773246} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.509832] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.510066] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Reconfigured VM to detach interface {{(pid=62600) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 879.740824] env[62600]: DEBUG oslo_vmware.api [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222595, 'name': PowerOnVM_Task, 'duration_secs': 0.429419} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.741312] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 879.748707] env[62600]: DEBUG nova.compute.manager [None req-f9e607d3-dd46-4692-bbd3-3853c91b15a5 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 879.749676] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-535949d2-a543-4fb9-bb14-ff8ada0ec4f0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.791499] env[62600]: DEBUG nova.compute.utils [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 879.837692] env[62600]: DEBUG nova.compute.utils [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 879.844291] env[62600]: DEBUG nova.compute.manager [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 879.844725] env[62600]: DEBUG nova.network.neutron [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 879.867403] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222596, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454379} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.867576] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 1b856aff-66be-4fa5-b8b7-124ac24fb4d1/1b856aff-66be-4fa5-b8b7-124ac24fb4d1.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 879.867810] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 879.868104] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e802fab9-5c79-41ba-bc74-012a019d8c18 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.881253] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 879.881253] env[62600]: value = "task-1222597" [ 879.881253] env[62600]: _type = "Task" [ 879.881253] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.892303] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222597, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.925781] env[62600]: DEBUG nova.policy [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '46e18992b34d412e8e3e8205d1d18265', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '734eda84fbe4493ba1f3db243a2bdb09', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 879.945822] env[62600]: DEBUG oslo_concurrency.lockutils [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.171674] env[62600]: DEBUG nova.compute.manager [req-07fd5823-a0ed-4c17-bb06-7f85e1cfa4f8 req-a9bd6c75-5ea9-4ae3-8314-4127733b5237 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Received event network-vif-deleted-a0ecf848-49ae-489e-9c07-0acfa758d69f {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.171865] env[62600]: INFO nova.compute.manager [req-07fd5823-a0ed-4c17-bb06-7f85e1cfa4f8 req-a9bd6c75-5ea9-4ae3-8314-4127733b5237 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Neutron deleted interface a0ecf848-49ae-489e-9c07-0acfa758d69f; detaching it from the instance and deleting it from the info cache [ 880.172194] env[62600]: DEBUG nova.network.neutron [req-07fd5823-a0ed-4c17-bb06-7f85e1cfa4f8 req-a9bd6c75-5ea9-4ae3-8314-4127733b5237 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Updating instance_info_cache with network_info: [{"id": "75e67d5c-8153-4a74-b875-19d68778a85a", "address": "fa:16:3e:fa:9d:b3", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75e67d5c-81", "ovs_interfaceid": "75e67d5c-8153-4a74-b875-19d68778a85a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.265835] env[62600]: DEBUG nova.network.neutron [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Successfully created port: f39debe1-7447-4681-a72a-83c77b315fd2 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 880.294344] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.352016] env[62600]: DEBUG nova.compute.manager [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 880.391889] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222597, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078686} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.395373] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 880.396524] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d60826b5-3ab8-4221-ab01-416f4e28a8e2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.420532] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 1b856aff-66be-4fa5-b8b7-124ac24fb4d1/1b856aff-66be-4fa5-b8b7-124ac24fb4d1.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 880.423804] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ef0e052-c6f1-405e-84bd-8daa1f414652 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.448121] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 880.448121] env[62600]: value = "task-1222598" [ 880.448121] env[62600]: _type = "Task" [ 880.448121] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.461206] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222598, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.667165] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e21693-8558-49cb-8c57-354908eed7b3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.675562] env[62600]: DEBUG oslo_concurrency.lockutils [req-07fd5823-a0ed-4c17-bb06-7f85e1cfa4f8 req-a9bd6c75-5ea9-4ae3-8314-4127733b5237 service nova] Acquiring lock "bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.676620] env[62600]: DEBUG oslo_concurrency.lockutils [req-07fd5823-a0ed-4c17-bb06-7f85e1cfa4f8 req-a9bd6c75-5ea9-4ae3-8314-4127733b5237 service nova] Acquired lock "bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.680886] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a3b5ca3-7f48-45fb-af8d-4d9c05310737 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.688400] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8c75a94-4f3c-4d89-b0f8-05d436d158b0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.712456] env[62600]: DEBUG oslo_concurrency.lockutils [req-07fd5823-a0ed-4c17-bb06-7f85e1cfa4f8 req-a9bd6c75-5ea9-4ae3-8314-4127733b5237 service nova] Releasing lock "bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.712784] env[62600]: WARNING nova.compute.manager [req-07fd5823-a0ed-4c17-bb06-7f85e1cfa4f8 req-a9bd6c75-5ea9-4ae3-8314-4127733b5237 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Detach interface failed, port_id=a0ecf848-49ae-489e-9c07-0acfa758d69f, reason: No device with interface-id a0ecf848-49ae-489e-9c07-0acfa758d69f exists on VM: nova.exception.NotFound: No device with interface-id a0ecf848-49ae-489e-9c07-0acfa758d69f exists on VM [ 880.742349] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c6641f2-77ee-49f3-a24f-11660aba81a2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.750955] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c536dcf4-7dac-4adf-8023-da2085a941c7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.766214] env[62600]: DEBUG nova.compute.provider_tree [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.941867] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "refresh_cache-bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.942252] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "refresh_cache-bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.942310] env[62600]: DEBUG nova.network.neutron [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 880.959612] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222598, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.271038] env[62600]: DEBUG nova.scheduler.client.report [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 881.332937] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "bd9b7351-01a2-429d-a860-d9ff6855eefc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.332937] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "bd9b7351-01a2-429d-a860-d9ff6855eefc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.332937] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "bd9b7351-01a2-429d-a860-d9ff6855eefc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.333085] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "bd9b7351-01a2-429d-a860-d9ff6855eefc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.333608] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "bd9b7351-01a2-429d-a860-d9ff6855eefc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.336348] env[62600]: INFO nova.compute.manager [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Terminating instance [ 881.338171] env[62600]: DEBUG nova.compute.manager [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 881.338360] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 881.339232] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a55424-afd4-45d8-966b-3638a9d4fc6e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.350141] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 881.350141] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-50474418-71fa-41ff-a2c2-96f18f958aa4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.359899] env[62600]: DEBUG oslo_vmware.api [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 881.359899] env[62600]: value = "task-1222599" [ 881.359899] env[62600]: _type = "Task" [ 881.359899] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.365165] env[62600]: DEBUG nova.compute.manager [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 881.379708] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquiring lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.380083] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.383094] env[62600]: INFO nova.compute.manager [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Attaching volume 83180b95-0343-4a55-9409-831dfcf80c02 to /dev/sdc [ 881.385334] env[62600]: DEBUG oslo_vmware.api [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222599, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.401248] env[62600]: DEBUG nova.virt.hardware [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='7be37b7e05f16cd8f7d40bf7b162ba75',container_format='bare',created_at=2024-10-09T12:04:47Z,direct_url=,disk_format='vmdk',id=e1395438-00bb-4e48-add8-ef3068f3c1c3,min_disk=1,min_ram=0,name='tempest-test-snap-664841953',owner='734eda84fbe4493ba1f3db243a2bdb09',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-09T12:05:05Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 881.401506] env[62600]: DEBUG nova.virt.hardware [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 881.401844] env[62600]: DEBUG nova.virt.hardware [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 881.402150] env[62600]: DEBUG nova.virt.hardware [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 881.402381] env[62600]: DEBUG nova.virt.hardware [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 881.403746] env[62600]: DEBUG nova.virt.hardware [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 881.404120] env[62600]: DEBUG nova.virt.hardware [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 881.404373] env[62600]: DEBUG nova.virt.hardware [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 881.404623] env[62600]: DEBUG nova.virt.hardware [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 881.404862] env[62600]: DEBUG nova.virt.hardware [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 881.405137] env[62600]: DEBUG nova.virt.hardware [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 881.406705] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cde8c26-0719-418c-9b6d-7b42e136b18b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.418673] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85aa0b52-e39d-4426-b7d0-c628a186aa34 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.438294] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f425d8f-a7fe-41f3-ac71-91738126d93f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.448771] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2067055-e7cd-48b2-abe0-e80bb33a188b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.459487] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222598, 'name': ReconfigVM_Task, 'duration_secs': 0.908847} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.459772] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 1b856aff-66be-4fa5-b8b7-124ac24fb4d1/1b856aff-66be-4fa5-b8b7-124ac24fb4d1.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 881.460414] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-31534e1a-ec0b-44a6-a930-5dc9c3c22f3d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.466818] env[62600]: DEBUG nova.virt.block_device [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Updating existing volume attachment record: acb4a91a-a362-4555-8ccb-a11142c4e2f2 {{(pid=62600) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 881.469992] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 881.469992] env[62600]: value = "task-1222600" [ 881.469992] env[62600]: _type = "Task" [ 881.469992] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.478273] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222600, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.777508] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.446s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.777508] env[62600]: DEBUG nova.compute.manager [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 881.780699] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.471s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.782453] env[62600]: INFO nova.compute.claims [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 881.873585] env[62600]: DEBUG oslo_vmware.api [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222599, 'name': PowerOffVM_Task, 'duration_secs': 0.238215} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.873953] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 881.874305] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 881.878020] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b60c7f89-2355-4066-a964-424f7f8fc296 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.900658] env[62600]: DEBUG nova.compute.manager [req-3d2be255-f0ef-43ba-9ecc-421603bb2c44 req-0e571b83-567c-4f87-ae64-7717001ffed8 service nova] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Received event network-vif-plugged-f39debe1-7447-4681-a72a-83c77b315fd2 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 881.900893] env[62600]: DEBUG oslo_concurrency.lockutils [req-3d2be255-f0ef-43ba-9ecc-421603bb2c44 req-0e571b83-567c-4f87-ae64-7717001ffed8 service nova] Acquiring lock "c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.901262] env[62600]: DEBUG oslo_concurrency.lockutils [req-3d2be255-f0ef-43ba-9ecc-421603bb2c44 req-0e571b83-567c-4f87-ae64-7717001ffed8 service nova] Lock "c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.901695] env[62600]: DEBUG oslo_concurrency.lockutils [req-3d2be255-f0ef-43ba-9ecc-421603bb2c44 req-0e571b83-567c-4f87-ae64-7717001ffed8 service nova] Lock "c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.901978] env[62600]: DEBUG nova.compute.manager [req-3d2be255-f0ef-43ba-9ecc-421603bb2c44 req-0e571b83-567c-4f87-ae64-7717001ffed8 service nova] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] No waiting events found dispatching network-vif-plugged-f39debe1-7447-4681-a72a-83c77b315fd2 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 881.902282] env[62600]: WARNING nova.compute.manager [req-3d2be255-f0ef-43ba-9ecc-421603bb2c44 req-0e571b83-567c-4f87-ae64-7717001ffed8 service nova] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Received unexpected event network-vif-plugged-f39debe1-7447-4681-a72a-83c77b315fd2 for instance with vm_state building and task_state spawning. [ 881.916656] env[62600]: DEBUG nova.network.neutron [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Updating instance_info_cache with network_info: [{"id": "75e67d5c-8153-4a74-b875-19d68778a85a", "address": "fa:16:3e:fa:9d:b3", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.198", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75e67d5c-81", "ovs_interfaceid": "75e67d5c-8153-4a74-b875-19d68778a85a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.955059] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 881.955382] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 881.955576] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Deleting the datastore file [datastore1] bd9b7351-01a2-429d-a860-d9ff6855eefc {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 881.955773] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ba80727-869b-49f0-abed-9e0066bfe27a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.964106] env[62600]: DEBUG oslo_vmware.api [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 881.964106] env[62600]: value = "task-1222603" [ 881.964106] env[62600]: _type = "Task" [ 881.964106] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.973881] env[62600]: DEBUG oslo_vmware.api [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222603, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.982941] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222600, 'name': Rename_Task, 'duration_secs': 0.148451} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.982941] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 881.986855] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9fd1902d-e894-4b10-be97-9ad175507ddf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.986855] env[62600]: INFO nova.compute.manager [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Unrescuing [ 881.986855] env[62600]: DEBUG oslo_concurrency.lockutils [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "refresh_cache-b967fb11-e70e-4e17-b769-38da581bd83b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.986855] env[62600]: DEBUG oslo_concurrency.lockutils [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquired lock "refresh_cache-b967fb11-e70e-4e17-b769-38da581bd83b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.986855] env[62600]: DEBUG nova.network.neutron [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 881.994018] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 881.994018] env[62600]: value = "task-1222604" [ 881.994018] env[62600]: _type = "Task" [ 881.994018] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.014072] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222604, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.070554] env[62600]: DEBUG nova.network.neutron [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Successfully updated port: f39debe1-7447-4681-a72a-83c77b315fd2 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 882.293710] env[62600]: DEBUG nova.compute.utils [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 882.296162] env[62600]: DEBUG nova.compute.manager [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 882.296437] env[62600]: DEBUG nova.network.neutron [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 882.378302] env[62600]: DEBUG nova.policy [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '94751040d0fd41a895010c0f2f019131', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef3bb131f3994f7a8673dafff6d353c1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 882.419027] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "refresh_cache-bd9b7351-01a2-429d-a860-d9ff6855eefc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.475584] env[62600]: DEBUG oslo_vmware.api [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222603, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16673} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.476430] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 882.476758] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 882.478202] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 882.478202] env[62600]: INFO nova.compute.manager [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Took 1.14 seconds to destroy the instance on the hypervisor. [ 882.478202] env[62600]: DEBUG oslo.service.loopingcall [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.478202] env[62600]: DEBUG nova.compute.manager [-] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 882.478202] env[62600]: DEBUG nova.network.neutron [-] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 882.505418] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222604, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.579603] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "refresh_cache-c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.580194] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquired lock "refresh_cache-c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.580499] env[62600]: DEBUG nova.network.neutron [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 882.807717] env[62600]: DEBUG nova.compute.manager [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 882.924678] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b11913af-60ed-454a-8c89-03f6a3c81306 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "interface-bd9b7351-01a2-429d-a860-d9ff6855eefc-a0ecf848-49ae-489e-9c07-0acfa758d69f" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.051s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.013027] env[62600]: DEBUG oslo_vmware.api [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222604, 'name': PowerOnVM_Task, 'duration_secs': 0.690288} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.013027] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 883.013027] env[62600]: INFO nova.compute.manager [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Took 8.53 seconds to spawn the instance on the hypervisor. [ 883.013027] env[62600]: DEBUG nova.compute.manager [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 883.013027] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0b227a2-0fd9-42b0-85de-97e7f6b8e107 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.020505] env[62600]: DEBUG nova.network.neutron [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Updating instance_info_cache with network_info: [{"id": "d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049", "address": "fa:16:3e:4f:fd:77", "network": {"id": "036dc114-6ea0-46cb-a958-4a6137e7e85f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-728627542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e1772c9a6c4671ab89df2e8b40cae1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6e6a2fb-7b", "ovs_interfaceid": "d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.080683] env[62600]: DEBUG nova.network.neutron [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Successfully created port: 517fe831-7f4c-477a-9c43-496ec9895bb3 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 883.127334] env[62600]: DEBUG nova.network.neutron [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 883.182933] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4066b490-6484-473a-99a7-0182395f2674 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.192909] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff6b0ec5-30cd-42ee-9f8e-ba8861538244 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.233364] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56b62ef-5de6-4d9b-98e2-6d44e0a5612a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.245814] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32aa1da0-3b03-4eaf-9249-755eb02f910f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.264604] env[62600]: DEBUG nova.compute.provider_tree [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.265576] env[62600]: DEBUG nova.compute.manager [req-7eef9bcf-4b8c-4519-88d2-7f685975d1c1 req-406c15f1-c7c8-4332-90db-a380565b5299 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Received event network-vif-deleted-75e67d5c-8153-4a74-b875-19d68778a85a {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.265753] env[62600]: INFO nova.compute.manager [req-7eef9bcf-4b8c-4519-88d2-7f685975d1c1 req-406c15f1-c7c8-4332-90db-a380565b5299 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Neutron deleted interface 75e67d5c-8153-4a74-b875-19d68778a85a; detaching it from the instance and deleting it from the info cache [ 883.266168] env[62600]: DEBUG nova.network.neutron [req-7eef9bcf-4b8c-4519-88d2-7f685975d1c1 req-406c15f1-c7c8-4332-90db-a380565b5299 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.381800] env[62600]: DEBUG nova.network.neutron [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Updating instance_info_cache with network_info: [{"id": "f39debe1-7447-4681-a72a-83c77b315fd2", "address": "fa:16:3e:c4:a5:5c", "network": {"id": "f59f04dc-029f-42be-92de-1d7d3e973fd7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-36149453-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "734eda84fbe4493ba1f3db243a2bdb09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb971244-43ba-41b4-a6a2-a4558548012c", "external-id": "nsx-vlan-transportzone-873", "segmentation_id": 873, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf39debe1-74", "ovs_interfaceid": "f39debe1-7447-4681-a72a-83c77b315fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.534888] env[62600]: DEBUG oslo_concurrency.lockutils [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Releasing lock "refresh_cache-b967fb11-e70e-4e17-b769-38da581bd83b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.535646] env[62600]: DEBUG nova.objects.instance [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lazy-loading 'flavor' on Instance uuid b967fb11-e70e-4e17-b769-38da581bd83b {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 883.545489] env[62600]: INFO nova.compute.manager [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Took 30.67 seconds to build instance. [ 883.736759] env[62600]: DEBUG nova.network.neutron [-] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.769065] env[62600]: DEBUG nova.scheduler.client.report [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 883.777193] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f9da827f-c2ed-4056-9f60-03ed70a16ff9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.786144] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-813ec126-4a5b-4a8f-a4eb-b1f1c340b119 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.822131] env[62600]: DEBUG nova.compute.manager [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 883.824368] env[62600]: DEBUG nova.compute.manager [req-7eef9bcf-4b8c-4519-88d2-7f685975d1c1 req-406c15f1-c7c8-4332-90db-a380565b5299 service nova] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Detach interface failed, port_id=75e67d5c-8153-4a74-b875-19d68778a85a, reason: Instance bd9b7351-01a2-429d-a860-d9ff6855eefc could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 883.851087] env[62600]: DEBUG nova.virt.hardware [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 883.851314] env[62600]: DEBUG nova.virt.hardware [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 883.851485] env[62600]: DEBUG nova.virt.hardware [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 883.851673] env[62600]: DEBUG nova.virt.hardware [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 883.851823] env[62600]: DEBUG nova.virt.hardware [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 883.851975] env[62600]: DEBUG nova.virt.hardware [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 883.852250] env[62600]: DEBUG nova.virt.hardware [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 883.852425] env[62600]: DEBUG nova.virt.hardware [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 883.852619] env[62600]: DEBUG nova.virt.hardware [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 883.853026] env[62600]: DEBUG nova.virt.hardware [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 883.853026] env[62600]: DEBUG nova.virt.hardware [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 883.853841] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56dc95aa-0692-48a4-bb82-ecd44a0d7b0c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.862382] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-997986a5-1084-4cd6-b243-617198b388fc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.884390] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Releasing lock "refresh_cache-c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.884996] env[62600]: DEBUG nova.compute.manager [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Instance network_info: |[{"id": "f39debe1-7447-4681-a72a-83c77b315fd2", "address": "fa:16:3e:c4:a5:5c", "network": {"id": "f59f04dc-029f-42be-92de-1d7d3e973fd7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-36149453-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "734eda84fbe4493ba1f3db243a2bdb09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb971244-43ba-41b4-a6a2-a4558548012c", "external-id": "nsx-vlan-transportzone-873", "segmentation_id": 873, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf39debe1-74", "ovs_interfaceid": "f39debe1-7447-4681-a72a-83c77b315fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 883.885353] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:a5:5c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cb971244-43ba-41b4-a6a2-a4558548012c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f39debe1-7447-4681-a72a-83c77b315fd2', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 883.893218] env[62600]: DEBUG oslo.service.loopingcall [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.893444] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 883.893671] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-607938bd-41b0-433c-b27b-83bf386bce09 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.915051] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 883.915051] env[62600]: value = "task-1222606" [ 883.915051] env[62600]: _type = "Task" [ 883.915051] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.925426] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222606, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.047209] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d45aef89-95fd-415a-9d4c-e982b354ac33 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "1b856aff-66be-4fa5-b8b7-124ac24fb4d1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.182s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.049686] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-894b85ae-30cb-49e8-b1ee-97c9e41054e5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.074238] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 884.074424] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1f05ec76-a869-4d90-a69a-8247590a7ca3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.083066] env[62600]: DEBUG oslo_vmware.api [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 884.083066] env[62600]: value = "task-1222607" [ 884.083066] env[62600]: _type = "Task" [ 884.083066] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.092414] env[62600]: DEBUG oslo_vmware.api [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222607, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.104449] env[62600]: DEBUG nova.compute.manager [req-50311446-6c64-488a-b6e2-9f4438a74b26 req-ea8b4d6e-109b-4fb5-9d33-35b785abfb7f service nova] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Received event network-changed-f39debe1-7447-4681-a72a-83c77b315fd2 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 884.104449] env[62600]: DEBUG nova.compute.manager [req-50311446-6c64-488a-b6e2-9f4438a74b26 req-ea8b4d6e-109b-4fb5-9d33-35b785abfb7f service nova] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Refreshing instance network info cache due to event network-changed-f39debe1-7447-4681-a72a-83c77b315fd2. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 884.104449] env[62600]: DEBUG oslo_concurrency.lockutils [req-50311446-6c64-488a-b6e2-9f4438a74b26 req-ea8b4d6e-109b-4fb5-9d33-35b785abfb7f service nova] Acquiring lock "refresh_cache-c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.104449] env[62600]: DEBUG oslo_concurrency.lockutils [req-50311446-6c64-488a-b6e2-9f4438a74b26 req-ea8b4d6e-109b-4fb5-9d33-35b785abfb7f service nova] Acquired lock "refresh_cache-c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.104449] env[62600]: DEBUG nova.network.neutron [req-50311446-6c64-488a-b6e2-9f4438a74b26 req-ea8b4d6e-109b-4fb5-9d33-35b785abfb7f service nova] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Refreshing network info cache for port f39debe1-7447-4681-a72a-83c77b315fd2 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 884.243331] env[62600]: INFO nova.compute.manager [-] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Took 1.77 seconds to deallocate network for instance. [ 884.267775] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Acquiring lock "9ed08410-a4cb-43c4-9271-2c04d9c87eac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.268589] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Lock "9ed08410-a4cb-43c4-9271-2c04d9c87eac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.276528] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.496s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.277283] env[62600]: DEBUG nova.compute.manager [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 884.280254] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.882s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.282351] env[62600]: INFO nova.compute.claims [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 884.425936] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222606, 'name': CreateVM_Task, 'duration_secs': 0.347001} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.426127] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 884.426897] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e1395438-00bb-4e48-add8-ef3068f3c1c3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.427108] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e1395438-00bb-4e48-add8-ef3068f3c1c3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.427448] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e1395438-00bb-4e48-add8-ef3068f3c1c3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 884.427715] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b02d49f-8fd4-41d7-a19c-730e90f432b9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.433057] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 884.433057] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52606dc9-37a2-b4f4-8b49-542256b6af69" [ 884.433057] env[62600]: _type = "Task" [ 884.433057] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.445321] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52606dc9-37a2-b4f4-8b49-542256b6af69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.596324] env[62600]: DEBUG oslo_vmware.api [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222607, 'name': PowerOffVM_Task, 'duration_secs': 0.210664} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.596695] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 884.602766] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Reconfiguring VM instance instance-00000041 to detach disk 2001 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 884.604692] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d77cd802-f574-4a5d-8a11-f3551e37f12f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.629956] env[62600]: DEBUG oslo_vmware.api [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 884.629956] env[62600]: value = "task-1222608" [ 884.629956] env[62600]: _type = "Task" [ 884.629956] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.643354] env[62600]: DEBUG oslo_vmware.api [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222608, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.752290] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.776047] env[62600]: DEBUG nova.compute.manager [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 884.789216] env[62600]: DEBUG nova.compute.utils [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 884.792472] env[62600]: DEBUG nova.compute.manager [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 884.792658] env[62600]: DEBUG nova.network.neutron [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 884.856323] env[62600]: DEBUG nova.policy [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'edbeccb7a21c465aa97117d8228d018f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ec4174c333ce49efb0805dea64bc343f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 884.915428] env[62600]: DEBUG nova.network.neutron [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Successfully updated port: 517fe831-7f4c-477a-9c43-496ec9895bb3 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 884.952585] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e1395438-00bb-4e48-add8-ef3068f3c1c3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.952898] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Processing image e1395438-00bb-4e48-add8-ef3068f3c1c3 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 884.953121] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e1395438-00bb-4e48-add8-ef3068f3c1c3/e1395438-00bb-4e48-add8-ef3068f3c1c3.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.953279] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e1395438-00bb-4e48-add8-ef3068f3c1c3/e1395438-00bb-4e48-add8-ef3068f3c1c3.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.953505] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 884.953742] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cb962f90-1728-44bc-a1a2-b1555d3b0ac9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.968497] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 884.968497] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 884.969886] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffb6f8ca-6d6f-4b72-9532-a803b2441df4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.973657] env[62600]: DEBUG nova.network.neutron [req-50311446-6c64-488a-b6e2-9f4438a74b26 req-ea8b4d6e-109b-4fb5-9d33-35b785abfb7f service nova] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Updated VIF entry in instance network info cache for port f39debe1-7447-4681-a72a-83c77b315fd2. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 884.973657] env[62600]: DEBUG nova.network.neutron [req-50311446-6c64-488a-b6e2-9f4438a74b26 req-ea8b4d6e-109b-4fb5-9d33-35b785abfb7f service nova] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Updating instance_info_cache with network_info: [{"id": "f39debe1-7447-4681-a72a-83c77b315fd2", "address": "fa:16:3e:c4:a5:5c", "network": {"id": "f59f04dc-029f-42be-92de-1d7d3e973fd7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-36149453-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "734eda84fbe4493ba1f3db243a2bdb09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb971244-43ba-41b4-a6a2-a4558548012c", "external-id": "nsx-vlan-transportzone-873", "segmentation_id": 873, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf39debe1-74", "ovs_interfaceid": "f39debe1-7447-4681-a72a-83c77b315fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.978599] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 884.978599] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525bf7c5-d2fe-30cb-08cb-ac0e70391ad0" [ 884.978599] env[62600]: _type = "Task" [ 884.978599] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.991594] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525bf7c5-d2fe-30cb-08cb-ac0e70391ad0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.140173] env[62600]: DEBUG oslo_vmware.api [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222608, 'name': ReconfigVM_Task, 'duration_secs': 0.371486} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.140484] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Reconfigured VM instance instance-00000041 to detach disk 2001 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 885.140689] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 885.140952] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8ae42af7-2249-4888-a98e-11f1133e0890 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.147736] env[62600]: DEBUG oslo_vmware.api [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 885.147736] env[62600]: value = "task-1222609" [ 885.147736] env[62600]: _type = "Task" [ 885.147736] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.164910] env[62600]: DEBUG oslo_vmware.api [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222609, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.224865] env[62600]: DEBUG nova.network.neutron [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Successfully created port: 99feed2a-122d-4f5d-890e-0f15c8614b88 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 885.294088] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.296058] env[62600]: DEBUG nova.compute.manager [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 885.419531] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "refresh_cache-ab11cdd1-80a7-43c4-afac-24c7930238c1" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.419698] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquired lock "refresh_cache-ab11cdd1-80a7-43c4-afac-24c7930238c1" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.419849] env[62600]: DEBUG nova.network.neutron [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 885.439375] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "6447a26d-fc30-41eb-bd42-94341a86c2e8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.439782] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "6447a26d-fc30-41eb-bd42-94341a86c2e8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.476604] env[62600]: DEBUG oslo_concurrency.lockutils [req-50311446-6c64-488a-b6e2-9f4438a74b26 req-ea8b4d6e-109b-4fb5-9d33-35b785abfb7f service nova] Releasing lock "refresh_cache-c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.493022] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Preparing fetch location {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 885.493473] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Fetch image to [datastore2] OSTACK_IMG_87d640a2-b3e2-4e57-8010-cf5090eaa846/OSTACK_IMG_87d640a2-b3e2-4e57-8010-cf5090eaa846.vmdk {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 885.493666] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Downloading stream optimized image e1395438-00bb-4e48-add8-ef3068f3c1c3 to [datastore2] OSTACK_IMG_87d640a2-b3e2-4e57-8010-cf5090eaa846/OSTACK_IMG_87d640a2-b3e2-4e57-8010-cf5090eaa846.vmdk on the data store datastore2 as vApp {{(pid=62600) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 885.493861] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Downloading image file data e1395438-00bb-4e48-add8-ef3068f3c1c3 to the ESX as VM named 'OSTACK_IMG_87d640a2-b3e2-4e57-8010-cf5090eaa846' {{(pid=62600) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 885.597853] env[62600]: DEBUG oslo_vmware.rw_handles [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 885.597853] env[62600]: value = "resgroup-9" [ 885.597853] env[62600]: _type = "ResourcePool" [ 885.597853] env[62600]: }. {{(pid=62600) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 885.598342] env[62600]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-bc58809c-e3ef-4d1d-afa7-11af4997ac75 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.624477] env[62600]: DEBUG oslo_vmware.rw_handles [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lease: (returnval){ [ 885.624477] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52637c69-f372-bd8f-2c43-8486772bef07" [ 885.624477] env[62600]: _type = "HttpNfcLease" [ 885.624477] env[62600]: } obtained for vApp import into resource pool (val){ [ 885.624477] env[62600]: value = "resgroup-9" [ 885.624477] env[62600]: _type = "ResourcePool" [ 885.624477] env[62600]: }. {{(pid=62600) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 885.624863] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the lease: (returnval){ [ 885.624863] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52637c69-f372-bd8f-2c43-8486772bef07" [ 885.624863] env[62600]: _type = "HttpNfcLease" [ 885.624863] env[62600]: } to be ready. {{(pid=62600) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 885.632909] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 885.632909] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52637c69-f372-bd8f-2c43-8486772bef07" [ 885.632909] env[62600]: _type = "HttpNfcLease" [ 885.632909] env[62600]: } is initializing. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 885.655813] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-382b13fb-857f-49f7-aaee-dc9ac5df8603 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.677608] env[62600]: DEBUG oslo_vmware.api [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222609, 'name': PowerOnVM_Task, 'duration_secs': 0.454117} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.678629] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 885.678925] env[62600]: DEBUG nova.compute.manager [None req-49eae920-e0af-4373-82c7-1772694f3286 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 885.683035] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd3eb65-a5e4-4a5a-bc3d-b20bdea083e8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.685544] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a6fccc-5727-4b66-9243-10256ae56146 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.732601] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b6cdc8-525b-462b-8c18-526c993bdb7d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.742175] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b9654a8-b301-460d-8740-96c15a695068 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.758537] env[62600]: DEBUG nova.compute.provider_tree [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.961052] env[62600]: DEBUG nova.network.neutron [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 886.027097] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Volume attach. Driver type: vmdk {{(pid=62600) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 886.027412] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264303', 'volume_id': '83180b95-0343-4a55-9409-831dfcf80c02', 'name': 'volume-83180b95-0343-4a55-9409-831dfcf80c02', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dde6df10-618a-40a8-b33f-efc0ca3a9287', 'attached_at': '', 'detached_at': '', 'volume_id': '83180b95-0343-4a55-9409-831dfcf80c02', 'serial': '83180b95-0343-4a55-9409-831dfcf80c02'} {{(pid=62600) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 886.028285] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a2968ce-ecce-4ce7-b3b0-040e053fedb5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.045742] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e93f0ef4-41f9-436f-b233-d97827bc852c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.073049] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] volume-83180b95-0343-4a55-9409-831dfcf80c02/volume-83180b95-0343-4a55-9409-831dfcf80c02.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.075599] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89cbe9ba-367a-41b8-a7a1-f1f9ab5acada {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.098256] env[62600]: DEBUG oslo_vmware.api [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for the task: (returnval){ [ 886.098256] env[62600]: value = "task-1222611" [ 886.098256] env[62600]: _type = "Task" [ 886.098256] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.107630] env[62600]: DEBUG oslo_vmware.api [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222611, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.130401] env[62600]: DEBUG nova.network.neutron [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Updating instance_info_cache with network_info: [{"id": "517fe831-7f4c-477a-9c43-496ec9895bb3", "address": "fa:16:3e:74:95:a0", "network": {"id": "e7547ac4-24a6-4778-a17a-40e79e7a0d0c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-320671681-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3bb131f3994f7a8673dafff6d353c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap517fe831-7f", "ovs_interfaceid": "517fe831-7f4c-477a-9c43-496ec9895bb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.133810] env[62600]: DEBUG nova.compute.manager [req-27690e1a-9b5c-409f-977e-723135e993ff req-380059d7-704c-4e87-a308-2648d50b0f7e service nova] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Received event network-vif-plugged-517fe831-7f4c-477a-9c43-496ec9895bb3 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.134080] env[62600]: DEBUG oslo_concurrency.lockutils [req-27690e1a-9b5c-409f-977e-723135e993ff req-380059d7-704c-4e87-a308-2648d50b0f7e service nova] Acquiring lock "ab11cdd1-80a7-43c4-afac-24c7930238c1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.134338] env[62600]: DEBUG oslo_concurrency.lockutils [req-27690e1a-9b5c-409f-977e-723135e993ff req-380059d7-704c-4e87-a308-2648d50b0f7e service nova] Lock "ab11cdd1-80a7-43c4-afac-24c7930238c1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.134606] env[62600]: DEBUG oslo_concurrency.lockutils [req-27690e1a-9b5c-409f-977e-723135e993ff req-380059d7-704c-4e87-a308-2648d50b0f7e service nova] Lock "ab11cdd1-80a7-43c4-afac-24c7930238c1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.134887] env[62600]: DEBUG nova.compute.manager [req-27690e1a-9b5c-409f-977e-723135e993ff req-380059d7-704c-4e87-a308-2648d50b0f7e service nova] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] No waiting events found dispatching network-vif-plugged-517fe831-7f4c-477a-9c43-496ec9895bb3 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 886.135183] env[62600]: WARNING nova.compute.manager [req-27690e1a-9b5c-409f-977e-723135e993ff req-380059d7-704c-4e87-a308-2648d50b0f7e service nova] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Received unexpected event network-vif-plugged-517fe831-7f4c-477a-9c43-496ec9895bb3 for instance with vm_state building and task_state spawning. [ 886.135428] env[62600]: DEBUG nova.compute.manager [req-27690e1a-9b5c-409f-977e-723135e993ff req-380059d7-704c-4e87-a308-2648d50b0f7e service nova] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Received event network-changed-517fe831-7f4c-477a-9c43-496ec9895bb3 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.135617] env[62600]: DEBUG nova.compute.manager [req-27690e1a-9b5c-409f-977e-723135e993ff req-380059d7-704c-4e87-a308-2648d50b0f7e service nova] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Refreshing instance network info cache due to event network-changed-517fe831-7f4c-477a-9c43-496ec9895bb3. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 886.135795] env[62600]: DEBUG oslo_concurrency.lockutils [req-27690e1a-9b5c-409f-977e-723135e993ff req-380059d7-704c-4e87-a308-2648d50b0f7e service nova] Acquiring lock "refresh_cache-ab11cdd1-80a7-43c4-afac-24c7930238c1" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.144134] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 886.144134] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52637c69-f372-bd8f-2c43-8486772bef07" [ 886.144134] env[62600]: _type = "HttpNfcLease" [ 886.144134] env[62600]: } is initializing. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 886.261656] env[62600]: DEBUG nova.scheduler.client.report [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.309965] env[62600]: DEBUG nova.compute.manager [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 886.344604] env[62600]: DEBUG nova.virt.hardware [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 886.344857] env[62600]: DEBUG nova.virt.hardware [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 886.345033] env[62600]: DEBUG nova.virt.hardware [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 886.345228] env[62600]: DEBUG nova.virt.hardware [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 886.345381] env[62600]: DEBUG nova.virt.hardware [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 886.345529] env[62600]: DEBUG nova.virt.hardware [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 886.345736] env[62600]: DEBUG nova.virt.hardware [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 886.345898] env[62600]: DEBUG nova.virt.hardware [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 886.346079] env[62600]: DEBUG nova.virt.hardware [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 886.346250] env[62600]: DEBUG nova.virt.hardware [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 886.346431] env[62600]: DEBUG nova.virt.hardware [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 886.347311] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6f34624-81b4-4f25-b703-77ec0c95d23c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.356051] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51917e83-f79f-413f-be5f-28520d16f0ba {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.608745] env[62600]: DEBUG oslo_vmware.api [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222611, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.634694] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 886.634694] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52637c69-f372-bd8f-2c43-8486772bef07" [ 886.634694] env[62600]: _type = "HttpNfcLease" [ 886.634694] env[62600]: } is ready. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 886.634998] env[62600]: DEBUG oslo_vmware.rw_handles [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 886.634998] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52637c69-f372-bd8f-2c43-8486772bef07" [ 886.634998] env[62600]: _type = "HttpNfcLease" [ 886.634998] env[62600]: }. {{(pid=62600) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 886.635773] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9be450-a239-49c3-a599-7b0b179bd331 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.638510] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Releasing lock "refresh_cache-ab11cdd1-80a7-43c4-afac-24c7930238c1" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.638815] env[62600]: DEBUG nova.compute.manager [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Instance network_info: |[{"id": "517fe831-7f4c-477a-9c43-496ec9895bb3", "address": "fa:16:3e:74:95:a0", "network": {"id": "e7547ac4-24a6-4778-a17a-40e79e7a0d0c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-320671681-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3bb131f3994f7a8673dafff6d353c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap517fe831-7f", "ovs_interfaceid": "517fe831-7f4c-477a-9c43-496ec9895bb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 886.639129] env[62600]: DEBUG oslo_concurrency.lockutils [req-27690e1a-9b5c-409f-977e-723135e993ff req-380059d7-704c-4e87-a308-2648d50b0f7e service nova] Acquired lock "refresh_cache-ab11cdd1-80a7-43c4-afac-24c7930238c1" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.639313] env[62600]: DEBUG nova.network.neutron [req-27690e1a-9b5c-409f-977e-723135e993ff req-380059d7-704c-4e87-a308-2648d50b0f7e service nova] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Refreshing network info cache for port 517fe831-7f4c-477a-9c43-496ec9895bb3 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 886.640468] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:74:95:a0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '517fe831-7f4c-477a-9c43-496ec9895bb3', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 886.648312] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Creating folder: Project (ef3bb131f3994f7a8673dafff6d353c1). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 886.649682] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-42687af7-a3b9-4e0c-af0b-341e671cb33f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.654945] env[62600]: DEBUG oslo_vmware.rw_handles [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef0503-6102-7169-86c9-b00ea9e75a98/disk-0.vmdk from lease info. {{(pid=62600) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 886.655157] env[62600]: DEBUG oslo_vmware.rw_handles [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef0503-6102-7169-86c9-b00ea9e75a98/disk-0.vmdk. {{(pid=62600) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 886.715463] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Created folder: Project (ef3bb131f3994f7a8673dafff6d353c1) in parent group-v264198. [ 886.715667] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Creating folder: Instances. Parent ref: group-v264306. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 886.715918] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4f0ec534-a520-4d88-b4db-256cb200f955 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.724361] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-23e9728e-765c-4dda-a92e-3d4a4f4691fa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.731407] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Created folder: Instances in parent group-v264306. [ 886.731667] env[62600]: DEBUG oslo.service.loopingcall [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 886.732493] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 886.732791] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-75686a23-741d-4f97-a6c0-8174b2934d5e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.755238] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 886.755238] env[62600]: value = "task-1222614" [ 886.755238] env[62600]: _type = "Task" [ 886.755238] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.764383] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222614, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.766376] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.766929] env[62600]: DEBUG nova.compute.manager [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 886.769772] env[62600]: DEBUG oslo_concurrency.lockutils [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.339s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.771413] env[62600]: INFO nova.compute.claims [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 886.942713] env[62600]: DEBUG nova.compute.manager [req-4399220d-6548-4f86-9193-1fb4a8fe753a req-8d7b72d9-ddc0-42be-ae57-1610a5caabb4 service nova] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Received event network-vif-plugged-99feed2a-122d-4f5d-890e-0f15c8614b88 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.942960] env[62600]: DEBUG oslo_concurrency.lockutils [req-4399220d-6548-4f86-9193-1fb4a8fe753a req-8d7b72d9-ddc0-42be-ae57-1610a5caabb4 service nova] Acquiring lock "f6957b5a-5da8-4205-92e5-2f08bf948e88-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.943766] env[62600]: DEBUG oslo_concurrency.lockutils [req-4399220d-6548-4f86-9193-1fb4a8fe753a req-8d7b72d9-ddc0-42be-ae57-1610a5caabb4 service nova] Lock "f6957b5a-5da8-4205-92e5-2f08bf948e88-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.943967] env[62600]: DEBUG oslo_concurrency.lockutils [req-4399220d-6548-4f86-9193-1fb4a8fe753a req-8d7b72d9-ddc0-42be-ae57-1610a5caabb4 service nova] Lock "f6957b5a-5da8-4205-92e5-2f08bf948e88-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.944176] env[62600]: DEBUG nova.compute.manager [req-4399220d-6548-4f86-9193-1fb4a8fe753a req-8d7b72d9-ddc0-42be-ae57-1610a5caabb4 service nova] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] No waiting events found dispatching network-vif-plugged-99feed2a-122d-4f5d-890e-0f15c8614b88 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 886.944345] env[62600]: WARNING nova.compute.manager [req-4399220d-6548-4f86-9193-1fb4a8fe753a req-8d7b72d9-ddc0-42be-ae57-1610a5caabb4 service nova] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Received unexpected event network-vif-plugged-99feed2a-122d-4f5d-890e-0f15c8614b88 for instance with vm_state building and task_state spawning. [ 887.116704] env[62600]: DEBUG oslo_vmware.api [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222611, 'name': ReconfigVM_Task, 'duration_secs': 0.880803} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.117204] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Reconfigured VM instance instance-0000003b to attach disk [datastore2] volume-83180b95-0343-4a55-9409-831dfcf80c02/volume-83180b95-0343-4a55-9409-831dfcf80c02.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.122971] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10ed60ec-3fc2-485d-9167-964ff476c0dc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.139288] env[62600]: DEBUG nova.network.neutron [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Successfully updated port: 99feed2a-122d-4f5d-890e-0f15c8614b88 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 887.155647] env[62600]: DEBUG oslo_vmware.api [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for the task: (returnval){ [ 887.155647] env[62600]: value = "task-1222615" [ 887.155647] env[62600]: _type = "Task" [ 887.155647] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.173349] env[62600]: DEBUG oslo_vmware.api [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222615, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.267337] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222614, 'name': CreateVM_Task, 'duration_secs': 0.408406} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.272573] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 887.273458] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.273674] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.274017] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 887.274906] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ea688bb-35f3-404d-8064-3104f812741a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.277438] env[62600]: DEBUG nova.compute.utils [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 887.283574] env[62600]: DEBUG nova.compute.manager [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 887.283574] env[62600]: DEBUG nova.network.neutron [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 887.289013] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 887.289013] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f6935a-481f-4285-1893-9b10e6928d42" [ 887.289013] env[62600]: _type = "Task" [ 887.289013] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.299567] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f6935a-481f-4285-1893-9b10e6928d42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.336090] env[62600]: DEBUG nova.policy [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b24ebe18c1444edaa8000d83e5c4ff6d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '888577fb598f45c881757ce3ee52bfa3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 887.521364] env[62600]: DEBUG nova.network.neutron [req-27690e1a-9b5c-409f-977e-723135e993ff req-380059d7-704c-4e87-a308-2648d50b0f7e service nova] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Updated VIF entry in instance network info cache for port 517fe831-7f4c-477a-9c43-496ec9895bb3. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 887.521793] env[62600]: DEBUG nova.network.neutron [req-27690e1a-9b5c-409f-977e-723135e993ff req-380059d7-704c-4e87-a308-2648d50b0f7e service nova] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Updating instance_info_cache with network_info: [{"id": "517fe831-7f4c-477a-9c43-496ec9895bb3", "address": "fa:16:3e:74:95:a0", "network": {"id": "e7547ac4-24a6-4778-a17a-40e79e7a0d0c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-320671681-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3bb131f3994f7a8673dafff6d353c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap517fe831-7f", "ovs_interfaceid": "517fe831-7f4c-477a-9c43-496ec9895bb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.571666] env[62600]: DEBUG oslo_vmware.rw_handles [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Completed reading data from the image iterator. {{(pid=62600) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 887.571962] env[62600]: DEBUG oslo_vmware.rw_handles [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef0503-6102-7169-86c9-b00ea9e75a98/disk-0.vmdk. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 887.572984] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3021a4bb-a96c-4deb-8985-4f78fd7dc1ca {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.580353] env[62600]: DEBUG oslo_vmware.rw_handles [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef0503-6102-7169-86c9-b00ea9e75a98/disk-0.vmdk is in state: ready. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 887.580520] env[62600]: DEBUG oslo_vmware.rw_handles [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef0503-6102-7169-86c9-b00ea9e75a98/disk-0.vmdk. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 887.580750] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-6e74db64-c8b9-4bf1-a513-ebdfe92a3e7e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.630677] env[62600]: DEBUG nova.network.neutron [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Successfully created port: 420267c8-eb2f-4f1f-b041-8eca21b9118b {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 887.645365] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Acquiring lock "refresh_cache-f6957b5a-5da8-4205-92e5-2f08bf948e88" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.645414] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Acquired lock "refresh_cache-f6957b5a-5da8-4205-92e5-2f08bf948e88" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.645569] env[62600]: DEBUG nova.network.neutron [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 887.667934] env[62600]: DEBUG oslo_vmware.api [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222615, 'name': ReconfigVM_Task, 'duration_secs': 0.176689} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.668328] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264303', 'volume_id': '83180b95-0343-4a55-9409-831dfcf80c02', 'name': 'volume-83180b95-0343-4a55-9409-831dfcf80c02', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dde6df10-618a-40a8-b33f-efc0ca3a9287', 'attached_at': '', 'detached_at': '', 'volume_id': '83180b95-0343-4a55-9409-831dfcf80c02', 'serial': '83180b95-0343-4a55-9409-831dfcf80c02'} {{(pid=62600) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 887.767523] env[62600]: DEBUG oslo_vmware.rw_handles [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef0503-6102-7169-86c9-b00ea9e75a98/disk-0.vmdk. {{(pid=62600) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 887.767972] env[62600]: INFO nova.virt.vmwareapi.images [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Downloaded image file data e1395438-00bb-4e48-add8-ef3068f3c1c3 [ 887.768755] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a919f09d-9860-43d7-949b-0289f69f1541 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.785812] env[62600]: DEBUG nova.compute.manager [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 887.792706] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1f621b46-65a2-4fe6-a57c-54e2d44bd736 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.804456] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f6935a-481f-4285-1893-9b10e6928d42, 'name': SearchDatastore_Task, 'duration_secs': 0.020157} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.804787] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.805488] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 887.805488] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.805488] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.805805] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 887.805957] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-39a43b7e-1621-4bc6-a523-873ea9ced7bf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.816705] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 887.816928] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 887.817700] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a37343f3-69b2-49b8-88c5-382c2f1f402a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.825092] env[62600]: INFO nova.virt.vmwareapi.images [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] The imported VM was unregistered [ 887.827920] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Caching image {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 887.828257] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Creating directory with path [datastore2] devstack-image-cache_base/e1395438-00bb-4e48-add8-ef3068f3c1c3 {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 887.828548] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f31a8053-f7ba-496b-9cea-7b6159de9d00 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.832249] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 887.832249] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5229cceb-e6df-7111-ff5a-b9fe33dd22f6" [ 887.832249] env[62600]: _type = "Task" [ 887.832249] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.842549] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Created directory with path [datastore2] devstack-image-cache_base/e1395438-00bb-4e48-add8-ef3068f3c1c3 {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 887.842780] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_87d640a2-b3e2-4e57-8010-cf5090eaa846/OSTACK_IMG_87d640a2-b3e2-4e57-8010-cf5090eaa846.vmdk to [datastore2] devstack-image-cache_base/e1395438-00bb-4e48-add8-ef3068f3c1c3/e1395438-00bb-4e48-add8-ef3068f3c1c3.vmdk. {{(pid=62600) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 887.846211] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-c3874a46-1c25-4f3f-a678-7519ac668ecc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.848715] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5229cceb-e6df-7111-ff5a-b9fe33dd22f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.855463] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 887.855463] env[62600]: value = "task-1222617" [ 887.855463] env[62600]: _type = "Task" [ 887.855463] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.869718] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222617, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.027669] env[62600]: DEBUG oslo_concurrency.lockutils [req-27690e1a-9b5c-409f-977e-723135e993ff req-380059d7-704c-4e87-a308-2648d50b0f7e service nova] Releasing lock "refresh_cache-ab11cdd1-80a7-43c4-afac-24c7930238c1" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.088495] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6a65c66-3a3c-4d7f-8327-07eb4ef598b9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.098256] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3fa37e6-b267-4ae5-8880-a29dd29e357a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.134504] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9390419-dbdd-4949-b17e-dc78241e6cc7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.143749] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a17a7a1-b8bb-4693-8fce-5d741205d602 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.161976] env[62600]: DEBUG nova.compute.provider_tree [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.182686] env[62600]: DEBUG nova.network.neutron [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 888.345749] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5229cceb-e6df-7111-ff5a-b9fe33dd22f6, 'name': SearchDatastore_Task, 'duration_secs': 0.013171} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.346996] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7016941-fdc1-4852-9c6a-cc79b571bfc7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.352404] env[62600]: DEBUG nova.network.neutron [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Updating instance_info_cache with network_info: [{"id": "99feed2a-122d-4f5d-890e-0f15c8614b88", "address": "fa:16:3e:b4:7c:12", "network": {"id": "b656cd97-9788-4a9c-a0ab-15209df9cb39", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-2020296784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec4174c333ce49efb0805dea64bc343f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a1439ce-fa5c-420d-bcf0-083f4cc002cd", "external-id": "nsx-vlan-transportzone-413", "segmentation_id": 413, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99feed2a-12", "ovs_interfaceid": "99feed2a-122d-4f5d-890e-0f15c8614b88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.356085] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 888.356085] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f6a90b-27ad-6221-4855-155d6583b08a" [ 888.356085] env[62600]: _type = "Task" [ 888.356085] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.368118] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222617, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.372715] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f6a90b-27ad-6221-4855-155d6583b08a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.666818] env[62600]: DEBUG nova.scheduler.client.report [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 888.714732] env[62600]: DEBUG nova.objects.instance [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lazy-loading 'flavor' on Instance uuid dde6df10-618a-40a8-b33f-efc0ca3a9287 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 888.798036] env[62600]: DEBUG nova.compute.manager [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 888.826958] env[62600]: DEBUG nova.virt.hardware [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 888.827229] env[62600]: DEBUG nova.virt.hardware [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 888.827396] env[62600]: DEBUG nova.virt.hardware [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 888.827585] env[62600]: DEBUG nova.virt.hardware [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 888.827738] env[62600]: DEBUG nova.virt.hardware [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 888.827886] env[62600]: DEBUG nova.virt.hardware [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 888.828268] env[62600]: DEBUG nova.virt.hardware [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 888.828472] env[62600]: DEBUG nova.virt.hardware [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 888.828735] env[62600]: DEBUG nova.virt.hardware [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 888.828916] env[62600]: DEBUG nova.virt.hardware [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 888.829119] env[62600]: DEBUG nova.virt.hardware [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 888.830195] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a78d26-81c9-4d96-a931-6e0e46f23792 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.839916] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2991d2-80bb-41d4-a4b8-0bd64ad8f116 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.855988] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Releasing lock "refresh_cache-f6957b5a-5da8-4205-92e5-2f08bf948e88" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.856407] env[62600]: DEBUG nova.compute.manager [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Instance network_info: |[{"id": "99feed2a-122d-4f5d-890e-0f15c8614b88", "address": "fa:16:3e:b4:7c:12", "network": {"id": "b656cd97-9788-4a9c-a0ab-15209df9cb39", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-2020296784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec4174c333ce49efb0805dea64bc343f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a1439ce-fa5c-420d-bcf0-083f4cc002cd", "external-id": "nsx-vlan-transportzone-413", "segmentation_id": 413, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99feed2a-12", "ovs_interfaceid": "99feed2a-122d-4f5d-890e-0f15c8614b88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 888.857341] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b4:7c:12', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0a1439ce-fa5c-420d-bcf0-083f4cc002cd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '99feed2a-122d-4f5d-890e-0f15c8614b88', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 888.865911] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Creating folder: Project (ec4174c333ce49efb0805dea64bc343f). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 888.869904] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-19f08ee6-47cb-4527-9202-2753406d0e29 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.883712] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f6a90b-27ad-6221-4855-155d6583b08a, 'name': SearchDatastore_Task, 'duration_secs': 0.019755} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.885712] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.885819] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] ab11cdd1-80a7-43c4-afac-24c7930238c1/ab11cdd1-80a7-43c4-afac-24c7930238c1.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 888.886064] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222617, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.886304] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-de733a52-0073-4ce9-9400-495eb0fc4709 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.890442] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Created folder: Project (ec4174c333ce49efb0805dea64bc343f) in parent group-v264198. [ 888.890660] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Creating folder: Instances. Parent ref: group-v264309. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 888.891351] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6d4ec645-bbd3-4bdb-b000-6141ca9f897e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.896996] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 888.896996] env[62600]: value = "task-1222619" [ 888.896996] env[62600]: _type = "Task" [ 888.896996] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.902249] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Created folder: Instances in parent group-v264309. [ 888.902530] env[62600]: DEBUG oslo.service.loopingcall [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.903336] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 888.903336] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-33bb441d-5d0d-45c4-92a4-8570fff14c08 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.921617] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222619, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.928540] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 888.928540] env[62600]: value = "task-1222621" [ 888.928540] env[62600]: _type = "Task" [ 888.928540] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.938032] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222621, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.980943] env[62600]: DEBUG nova.compute.manager [req-940bfcbb-edf2-4867-9f11-874326eb130b req-7692d6e8-3443-4e99-8c0a-c8ab951475fa service nova] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Received event network-changed-99feed2a-122d-4f5d-890e-0f15c8614b88 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.980943] env[62600]: DEBUG nova.compute.manager [req-940bfcbb-edf2-4867-9f11-874326eb130b req-7692d6e8-3443-4e99-8c0a-c8ab951475fa service nova] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Refreshing instance network info cache due to event network-changed-99feed2a-122d-4f5d-890e-0f15c8614b88. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 888.981217] env[62600]: DEBUG oslo_concurrency.lockutils [req-940bfcbb-edf2-4867-9f11-874326eb130b req-7692d6e8-3443-4e99-8c0a-c8ab951475fa service nova] Acquiring lock "refresh_cache-f6957b5a-5da8-4205-92e5-2f08bf948e88" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.981419] env[62600]: DEBUG oslo_concurrency.lockutils [req-940bfcbb-edf2-4867-9f11-874326eb130b req-7692d6e8-3443-4e99-8c0a-c8ab951475fa service nova] Acquired lock "refresh_cache-f6957b5a-5da8-4205-92e5-2f08bf948e88" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.981619] env[62600]: DEBUG nova.network.neutron [req-940bfcbb-edf2-4867-9f11-874326eb130b req-7692d6e8-3443-4e99-8c0a-c8ab951475fa service nova] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Refreshing network info cache for port 99feed2a-122d-4f5d-890e-0f15c8614b88 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 889.176967] env[62600]: DEBUG oslo_concurrency.lockutils [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.407s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.178426] env[62600]: DEBUG nova.compute.manager [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 889.185221] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.678s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.185221] env[62600]: INFO nova.compute.claims [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 889.221613] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f7759c18-f90b-400d-a6bb-94463eea3077 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.839s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.315092] env[62600]: DEBUG nova.compute.manager [req-4b9ad3f5-b1d3-41bc-bac8-8bef370145fc req-fe4039c4-3f87-4791-8f24-1842b7ab252a service nova] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Received event network-vif-plugged-420267c8-eb2f-4f1f-b041-8eca21b9118b {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 889.315357] env[62600]: DEBUG oslo_concurrency.lockutils [req-4b9ad3f5-b1d3-41bc-bac8-8bef370145fc req-fe4039c4-3f87-4791-8f24-1842b7ab252a service nova] Acquiring lock "107ed783-10a3-4f07-a809-a5a076c2b904-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.315729] env[62600]: DEBUG oslo_concurrency.lockutils [req-4b9ad3f5-b1d3-41bc-bac8-8bef370145fc req-fe4039c4-3f87-4791-8f24-1842b7ab252a service nova] Lock "107ed783-10a3-4f07-a809-a5a076c2b904-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.316043] env[62600]: DEBUG oslo_concurrency.lockutils [req-4b9ad3f5-b1d3-41bc-bac8-8bef370145fc req-fe4039c4-3f87-4791-8f24-1842b7ab252a service nova] Lock "107ed783-10a3-4f07-a809-a5a076c2b904-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.316318] env[62600]: DEBUG nova.compute.manager [req-4b9ad3f5-b1d3-41bc-bac8-8bef370145fc req-fe4039c4-3f87-4791-8f24-1842b7ab252a service nova] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] No waiting events found dispatching network-vif-plugged-420267c8-eb2f-4f1f-b041-8eca21b9118b {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 889.316521] env[62600]: WARNING nova.compute.manager [req-4b9ad3f5-b1d3-41bc-bac8-8bef370145fc req-fe4039c4-3f87-4791-8f24-1842b7ab252a service nova] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Received unexpected event network-vif-plugged-420267c8-eb2f-4f1f-b041-8eca21b9118b for instance with vm_state building and task_state spawning. [ 889.377262] env[62600]: DEBUG nova.network.neutron [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Successfully updated port: 420267c8-eb2f-4f1f-b041-8eca21b9118b {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 889.385452] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222617, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.409423] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222619, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.442348] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222621, 'name': CreateVM_Task} progress is 25%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.690949] env[62600]: DEBUG nova.compute.utils [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 889.697403] env[62600]: DEBUG nova.compute.manager [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 889.697645] env[62600]: DEBUG nova.network.neutron [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 889.765160] env[62600]: DEBUG nova.network.neutron [req-940bfcbb-edf2-4867-9f11-874326eb130b req-7692d6e8-3443-4e99-8c0a-c8ab951475fa service nova] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Updated VIF entry in instance network info cache for port 99feed2a-122d-4f5d-890e-0f15c8614b88. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 889.765160] env[62600]: DEBUG nova.network.neutron [req-940bfcbb-edf2-4867-9f11-874326eb130b req-7692d6e8-3443-4e99-8c0a-c8ab951475fa service nova] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Updating instance_info_cache with network_info: [{"id": "99feed2a-122d-4f5d-890e-0f15c8614b88", "address": "fa:16:3e:b4:7c:12", "network": {"id": "b656cd97-9788-4a9c-a0ab-15209df9cb39", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-2020296784-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ec4174c333ce49efb0805dea64bc343f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0a1439ce-fa5c-420d-bcf0-083f4cc002cd", "external-id": "nsx-vlan-transportzone-413", "segmentation_id": 413, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99feed2a-12", "ovs_interfaceid": "99feed2a-122d-4f5d-890e-0f15c8614b88", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.772017] env[62600]: DEBUG nova.policy [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bd0d91fa2ac74c38b9e0ebf39c3baa7a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73e664d178f7484a9f4741b4d9450e68', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 889.885617] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquiring lock "refresh_cache-107ed783-10a3-4f07-a809-a5a076c2b904" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.886392] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquired lock "refresh_cache-107ed783-10a3-4f07-a809-a5a076c2b904" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.886805] env[62600]: DEBUG nova.network.neutron [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 889.888520] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222617, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.913024] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222619, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.672577} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.913024] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] ab11cdd1-80a7-43c4-afac-24c7930238c1/ab11cdd1-80a7-43c4-afac-24c7930238c1.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 889.913024] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 889.913024] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-686bb6df-85e0-476e-8eaf-c9c1b50536d2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.923834] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 889.923834] env[62600]: value = "task-1222622" [ 889.923834] env[62600]: _type = "Task" [ 889.923834] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.944713] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222621, 'name': CreateVM_Task, 'duration_secs': 0.732332} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.948638] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 889.948909] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222622, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.949721] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.950170] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.950613] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 889.950954] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc9ce9ea-ac44-4750-a92c-20a585869cb4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.958562] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Waiting for the task: (returnval){ [ 889.958562] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]524aac3a-fd1a-844a-bc93-7ecfaacc34c0" [ 889.958562] env[62600]: _type = "Task" [ 889.958562] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.973470] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]524aac3a-fd1a-844a-bc93-7ecfaacc34c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.201862] env[62600]: DEBUG nova.compute.manager [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 890.273469] env[62600]: DEBUG oslo_concurrency.lockutils [req-940bfcbb-edf2-4867-9f11-874326eb130b req-7692d6e8-3443-4e99-8c0a-c8ab951475fa service nova] Releasing lock "refresh_cache-f6957b5a-5da8-4205-92e5-2f08bf948e88" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.339620] env[62600]: DEBUG oslo_concurrency.lockutils [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "248a5371-6ff7-4da7-ae97-d638bea0123a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.339865] env[62600]: DEBUG oslo_concurrency.lockutils [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "248a5371-6ff7-4da7-ae97-d638bea0123a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.385104] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222617, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.441819} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.386022] env[62600]: INFO nova.virt.vmwareapi.ds_util [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_87d640a2-b3e2-4e57-8010-cf5090eaa846/OSTACK_IMG_87d640a2-b3e2-4e57-8010-cf5090eaa846.vmdk to [datastore2] devstack-image-cache_base/e1395438-00bb-4e48-add8-ef3068f3c1c3/e1395438-00bb-4e48-add8-ef3068f3c1c3.vmdk. [ 890.386154] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Cleaning up location [datastore2] OSTACK_IMG_87d640a2-b3e2-4e57-8010-cf5090eaa846 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 890.386433] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_87d640a2-b3e2-4e57-8010-cf5090eaa846 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 890.387094] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-476c380d-d295-4608-bec3-dd994763a7c7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.403370] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 890.403370] env[62600]: value = "task-1222623" [ 890.403370] env[62600]: _type = "Task" [ 890.403370] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.413190] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222623, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.421868] env[62600]: DEBUG nova.network.neutron [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 890.439435] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222622, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086184} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.442258] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 890.443172] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c4f5b60-97f6-49ca-9452-62a780d4dac7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.472451] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] ab11cdd1-80a7-43c4-afac-24c7930238c1/ab11cdd1-80a7-43c4-afac-24c7930238c1.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 890.481119] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e99947a-3c5b-4fb7-a9c9-09908661b5d6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.504072] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]524aac3a-fd1a-844a-bc93-7ecfaacc34c0, 'name': SearchDatastore_Task, 'duration_secs': 0.01676} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.505804] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.505804] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 890.506420] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.506420] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.506420] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 890.507179] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 890.507179] env[62600]: value = "task-1222624" [ 890.507179] env[62600]: _type = "Task" [ 890.507179] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.507179] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eaece7e3-6176-474f-a642-2019fe99afba {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.520521] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222624, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.521715] env[62600]: DEBUG nova.network.neutron [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Successfully created port: 92bc4652-839c-47b3-bcf5-2cc0c43d5537 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 890.524731] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 890.525672] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 890.528080] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24975922-4792-4888-9649-57ccd4a8b3f1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.538544] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Waiting for the task: (returnval){ [ 890.538544] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c10ff4-d2f7-f673-c581-5be6bba275b2" [ 890.538544] env[62600]: _type = "Task" [ 890.538544] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.547950] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c10ff4-d2f7-f673-c581-5be6bba275b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.565154] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da7b246-fdcf-4337-ae0b-2b0d2cc38243 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.572783] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f65687-02f0-43f7-b612-9294d9aa32d1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.602884] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5ddefda0-b3d6-4d04-afc4-3cf39aa5a1da tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquiring lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.602884] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5ddefda0-b3d6-4d04-afc4-3cf39aa5a1da tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.607371] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61616e23-c768-4b0b-9cab-01cf25212808 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.616534] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-355f4c14-354b-45a2-a3de-0fbf798135d5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.633206] env[62600]: DEBUG nova.compute.provider_tree [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.646110] env[62600]: DEBUG nova.network.neutron [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Updating instance_info_cache with network_info: [{"id": "420267c8-eb2f-4f1f-b041-8eca21b9118b", "address": "fa:16:3e:a3:28:03", "network": {"id": "102eb76e-a1ff-4b27-b285-516053dcad35", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-543226554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "888577fb598f45c881757ce3ee52bfa3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap420267c8-eb", "ovs_interfaceid": "420267c8-eb2f-4f1f-b041-8eca21b9118b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.915839] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222623, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.044607} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.916253] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 890.916372] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e1395438-00bb-4e48-add8-ef3068f3c1c3/e1395438-00bb-4e48-add8-ef3068f3c1c3.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.916542] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e1395438-00bb-4e48-add8-ef3068f3c1c3/e1395438-00bb-4e48-add8-ef3068f3c1c3.vmdk to [datastore2] c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f/c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 890.916808] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0626c574-eee1-46fa-b4e0-e1e454a3bae4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.924951] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 890.924951] env[62600]: value = "task-1222625" [ 890.924951] env[62600]: _type = "Task" [ 890.924951] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.933031] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222625, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.019797] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222624, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.049345] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c10ff4-d2f7-f673-c581-5be6bba275b2, 'name': SearchDatastore_Task, 'duration_secs': 0.011752} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.050228] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b518af1-947b-4094-86ac-78b194e09d09 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.056646] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Waiting for the task: (returnval){ [ 891.056646] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]527ca5f1-5b34-9414-1264-477f5d5da9b6" [ 891.056646] env[62600]: _type = "Task" [ 891.056646] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.066583] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]527ca5f1-5b34-9414-1264-477f5d5da9b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.111896] env[62600]: INFO nova.compute.manager [None req-5ddefda0-b3d6-4d04-afc4-3cf39aa5a1da tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Detaching volume 93edc5a4-1bb4-44c9-868a-43a307466fe0 [ 891.136207] env[62600]: DEBUG nova.scheduler.client.report [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 891.148248] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Releasing lock "refresh_cache-107ed783-10a3-4f07-a809-a5a076c2b904" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.148746] env[62600]: DEBUG nova.compute.manager [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Instance network_info: |[{"id": "420267c8-eb2f-4f1f-b041-8eca21b9118b", "address": "fa:16:3e:a3:28:03", "network": {"id": "102eb76e-a1ff-4b27-b285-516053dcad35", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-543226554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "888577fb598f45c881757ce3ee52bfa3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap420267c8-eb", "ovs_interfaceid": "420267c8-eb2f-4f1f-b041-8eca21b9118b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 891.149636] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a3:28:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3a0ddd7d-c321-4187-bdd8-b19044ea2c4a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '420267c8-eb2f-4f1f-b041-8eca21b9118b', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 891.158987] env[62600]: DEBUG oslo.service.loopingcall [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 891.161070] env[62600]: INFO nova.virt.block_device [None req-5ddefda0-b3d6-4d04-afc4-3cf39aa5a1da tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Attempting to driver detach volume 93edc5a4-1bb4-44c9-868a-43a307466fe0 from mountpoint /dev/sdb [ 891.161394] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ddefda0-b3d6-4d04-afc4-3cf39aa5a1da tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Volume detach. Driver type: vmdk {{(pid=62600) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 891.161640] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ddefda0-b3d6-4d04-afc4-3cf39aa5a1da tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264299', 'volume_id': '93edc5a4-1bb4-44c9-868a-43a307466fe0', 'name': 'volume-93edc5a4-1bb4-44c9-868a-43a307466fe0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dde6df10-618a-40a8-b33f-efc0ca3a9287', 'attached_at': '', 'detached_at': '', 'volume_id': '93edc5a4-1bb4-44c9-868a-43a307466fe0', 'serial': '93edc5a4-1bb4-44c9-868a-43a307466fe0'} {{(pid=62600) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 891.162319] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 891.163165] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87009205-41dd-4677-9501-a49aa0acb4da {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.166198] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-05304cc4-b978-4e82-93cf-5d7a943f68c3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.207797] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-076288df-46c9-4714-a9ff-80958925b0c2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.210518] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 891.210518] env[62600]: value = "task-1222626" [ 891.210518] env[62600]: _type = "Task" [ 891.210518] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.211797] env[62600]: DEBUG nova.compute.manager [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 891.223114] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dae9336b-effe-49d5-94e3-e424e7dbd425 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.229550] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222626, 'name': CreateVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.257077] env[62600]: DEBUG nova.virt.hardware [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 891.257300] env[62600]: DEBUG nova.virt.hardware [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 891.257437] env[62600]: DEBUG nova.virt.hardware [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 891.257603] env[62600]: DEBUG nova.virt.hardware [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 891.257754] env[62600]: DEBUG nova.virt.hardware [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 891.257909] env[62600]: DEBUG nova.virt.hardware [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 891.258143] env[62600]: DEBUG nova.virt.hardware [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 891.258312] env[62600]: DEBUG nova.virt.hardware [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 891.258485] env[62600]: DEBUG nova.virt.hardware [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 891.258652] env[62600]: DEBUG nova.virt.hardware [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 891.258829] env[62600]: DEBUG nova.virt.hardware [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 891.259795] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cef23fd3-db4f-4ab9-a549-43c948363e8a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.263231] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d8061f-e388-4148-b460-9304546e2100 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.283454] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06792879-6520-4cbf-b077-b3d623b07968 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.287025] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ddefda0-b3d6-4d04-afc4-3cf39aa5a1da tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] The volume has not been displaced from its original location: [datastore2] volume-93edc5a4-1bb4-44c9-868a-43a307466fe0/volume-93edc5a4-1bb4-44c9-868a-43a307466fe0.vmdk. No consolidation needed. {{(pid=62600) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 891.292707] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ddefda0-b3d6-4d04-afc4-3cf39aa5a1da tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Reconfiguring VM instance instance-0000003b to detach disk 2001 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 891.293516] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02ef78b2-f1b4-4da2-9737-146dd08c6f11 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.321381] env[62600]: DEBUG oslo_vmware.api [None req-5ddefda0-b3d6-4d04-afc4-3cf39aa5a1da tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for the task: (returnval){ [ 891.321381] env[62600]: value = "task-1222627" [ 891.321381] env[62600]: _type = "Task" [ 891.321381] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.331811] env[62600]: DEBUG oslo_vmware.api [None req-5ddefda0-b3d6-4d04-afc4-3cf39aa5a1da tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222627, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.398903] env[62600]: DEBUG nova.compute.manager [req-452709c4-a896-47cd-aba8-22bee99c95cd req-2e1da26a-5de8-4196-89dd-21c7f92f2802 service nova] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Received event network-changed-420267c8-eb2f-4f1f-b041-8eca21b9118b {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 891.399223] env[62600]: DEBUG nova.compute.manager [req-452709c4-a896-47cd-aba8-22bee99c95cd req-2e1da26a-5de8-4196-89dd-21c7f92f2802 service nova] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Refreshing instance network info cache due to event network-changed-420267c8-eb2f-4f1f-b041-8eca21b9118b. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 891.399492] env[62600]: DEBUG oslo_concurrency.lockutils [req-452709c4-a896-47cd-aba8-22bee99c95cd req-2e1da26a-5de8-4196-89dd-21c7f92f2802 service nova] Acquiring lock "refresh_cache-107ed783-10a3-4f07-a809-a5a076c2b904" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.399675] env[62600]: DEBUG oslo_concurrency.lockutils [req-452709c4-a896-47cd-aba8-22bee99c95cd req-2e1da26a-5de8-4196-89dd-21c7f92f2802 service nova] Acquired lock "refresh_cache-107ed783-10a3-4f07-a809-a5a076c2b904" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.399914] env[62600]: DEBUG nova.network.neutron [req-452709c4-a896-47cd-aba8-22bee99c95cd req-2e1da26a-5de8-4196-89dd-21c7f92f2802 service nova] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Refreshing network info cache for port 420267c8-eb2f-4f1f-b041-8eca21b9118b {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 891.436964] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222625, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.522569] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222624, 'name': ReconfigVM_Task, 'duration_secs': 0.895466} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.522881] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Reconfigured VM instance instance-00000046 to attach disk [datastore1] ab11cdd1-80a7-43c4-afac-24c7930238c1/ab11cdd1-80a7-43c4-afac-24c7930238c1.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.523726] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e3702b30-dc0c-4aa0-a4b4-8401fd3893e3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.532419] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 891.532419] env[62600]: value = "task-1222628" [ 891.532419] env[62600]: _type = "Task" [ 891.532419] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.552019] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222628, 'name': Rename_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.568481] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]527ca5f1-5b34-9414-1264-477f5d5da9b6, 'name': SearchDatastore_Task, 'duration_secs': 0.011311} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.569163] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.569163] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] f6957b5a-5da8-4205-92e5-2f08bf948e88/f6957b5a-5da8-4205-92e5-2f08bf948e88.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 891.569416] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-13a11810-ea0b-4e91-93db-efc407b464d1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.578343] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Waiting for the task: (returnval){ [ 891.578343] env[62600]: value = "task-1222629" [ 891.578343] env[62600]: _type = "Task" [ 891.578343] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.589335] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': task-1222629, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.641825] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.459s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.642559] env[62600]: DEBUG nova.compute.manager [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 891.645529] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.105s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.647827] env[62600]: INFO nova.compute.claims [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 891.722513] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222626, 'name': CreateVM_Task, 'duration_secs': 0.488108} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.722733] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 891.723550] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.723751] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.724109] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 891.724376] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87bcefe5-62ec-4c76-a52e-b402446ff5b0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.730351] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 891.730351] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52524e80-d338-1b07-2fa7-ae3911f1e9b1" [ 891.730351] env[62600]: _type = "Task" [ 891.730351] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.740142] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52524e80-d338-1b07-2fa7-ae3911f1e9b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.832534] env[62600]: DEBUG oslo_vmware.api [None req-5ddefda0-b3d6-4d04-afc4-3cf39aa5a1da tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222627, 'name': ReconfigVM_Task, 'duration_secs': 0.384012} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.832822] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ddefda0-b3d6-4d04-afc4-3cf39aa5a1da tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Reconfigured VM instance instance-0000003b to detach disk 2001 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 891.837870] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed173794-19c9-4b9d-93bb-ca908b915969 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.854550] env[62600]: DEBUG oslo_vmware.api [None req-5ddefda0-b3d6-4d04-afc4-3cf39aa5a1da tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for the task: (returnval){ [ 891.854550] env[62600]: value = "task-1222630" [ 891.854550] env[62600]: _type = "Task" [ 891.854550] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.865644] env[62600]: DEBUG oslo_vmware.api [None req-5ddefda0-b3d6-4d04-afc4-3cf39aa5a1da tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222630, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.938033] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222625, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.049121] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222628, 'name': Rename_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.089378] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': task-1222629, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.153992] env[62600]: DEBUG nova.compute.utils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 892.155447] env[62600]: DEBUG nova.compute.manager [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 892.155447] env[62600]: DEBUG nova.network.neutron [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 892.162910] env[62600]: DEBUG nova.network.neutron [req-452709c4-a896-47cd-aba8-22bee99c95cd req-2e1da26a-5de8-4196-89dd-21c7f92f2802 service nova] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Updated VIF entry in instance network info cache for port 420267c8-eb2f-4f1f-b041-8eca21b9118b. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 892.163687] env[62600]: DEBUG nova.network.neutron [req-452709c4-a896-47cd-aba8-22bee99c95cd req-2e1da26a-5de8-4196-89dd-21c7f92f2802 service nova] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Updating instance_info_cache with network_info: [{"id": "420267c8-eb2f-4f1f-b041-8eca21b9118b", "address": "fa:16:3e:a3:28:03", "network": {"id": "102eb76e-a1ff-4b27-b285-516053dcad35", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-543226554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "888577fb598f45c881757ce3ee52bfa3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap420267c8-eb", "ovs_interfaceid": "420267c8-eb2f-4f1f-b041-8eca21b9118b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.219451] env[62600]: DEBUG nova.policy [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '41f578fc709a497abd53d1a4f0540f6d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5e7eca5e197e47759eeea423e280986f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 892.241709] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52524e80-d338-1b07-2fa7-ae3911f1e9b1, 'name': SearchDatastore_Task, 'duration_secs': 0.015191} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.242047] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.242328] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 892.242647] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.242819] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.243021] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 892.243312] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-70cfb07c-7940-4f1d-8fc8-84637be9ac14 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.255337] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 892.255541] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 892.256338] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-416d2385-dbcb-4931-a170-647cdf0023e7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.263365] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 892.263365] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52819386-f9c8-df8c-8896-511a71627f88" [ 892.263365] env[62600]: _type = "Task" [ 892.263365] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.274893] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52819386-f9c8-df8c-8896-511a71627f88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.324979] env[62600]: DEBUG nova.compute.manager [req-678e51eb-60af-404b-95da-c843f1ab5e82 req-9bf4313e-6cac-4512-a83d-a63033edd6e2 service nova] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Received event network-vif-plugged-92bc4652-839c-47b3-bcf5-2cc0c43d5537 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.325239] env[62600]: DEBUG oslo_concurrency.lockutils [req-678e51eb-60af-404b-95da-c843f1ab5e82 req-9bf4313e-6cac-4512-a83d-a63033edd6e2 service nova] Acquiring lock "d627a701-77cd-4a1d-9e52-0fc4adc93391-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.325469] env[62600]: DEBUG oslo_concurrency.lockutils [req-678e51eb-60af-404b-95da-c843f1ab5e82 req-9bf4313e-6cac-4512-a83d-a63033edd6e2 service nova] Lock "d627a701-77cd-4a1d-9e52-0fc4adc93391-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.325651] env[62600]: DEBUG oslo_concurrency.lockutils [req-678e51eb-60af-404b-95da-c843f1ab5e82 req-9bf4313e-6cac-4512-a83d-a63033edd6e2 service nova] Lock "d627a701-77cd-4a1d-9e52-0fc4adc93391-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.325800] env[62600]: DEBUG nova.compute.manager [req-678e51eb-60af-404b-95da-c843f1ab5e82 req-9bf4313e-6cac-4512-a83d-a63033edd6e2 service nova] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] No waiting events found dispatching network-vif-plugged-92bc4652-839c-47b3-bcf5-2cc0c43d5537 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 892.326024] env[62600]: WARNING nova.compute.manager [req-678e51eb-60af-404b-95da-c843f1ab5e82 req-9bf4313e-6cac-4512-a83d-a63033edd6e2 service nova] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Received unexpected event network-vif-plugged-92bc4652-839c-47b3-bcf5-2cc0c43d5537 for instance with vm_state building and task_state spawning. [ 892.366168] env[62600]: DEBUG oslo_vmware.api [None req-5ddefda0-b3d6-4d04-afc4-3cf39aa5a1da tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222630, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.437383] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222625, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.500050] env[62600]: DEBUG nova.network.neutron [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Successfully updated port: 92bc4652-839c-47b3-bcf5-2cc0c43d5537 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 892.506151] env[62600]: DEBUG nova.network.neutron [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Successfully created port: e04d7519-77e4-4593-b7ec-5fc549cfce8f {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 892.545408] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222628, 'name': Rename_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.590769] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': task-1222629, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.673017} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.590966] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] f6957b5a-5da8-4205-92e5-2f08bf948e88/f6957b5a-5da8-4205-92e5-2f08bf948e88.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 892.591206] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 892.591615] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-97445989-6107-49ac-8811-9727f819b722 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.603609] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Waiting for the task: (returnval){ [ 892.603609] env[62600]: value = "task-1222631" [ 892.603609] env[62600]: _type = "Task" [ 892.603609] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.613364] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': task-1222631, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.664618] env[62600]: DEBUG nova.compute.manager [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 892.670030] env[62600]: DEBUG oslo_concurrency.lockutils [req-452709c4-a896-47cd-aba8-22bee99c95cd req-2e1da26a-5de8-4196-89dd-21c7f92f2802 service nova] Releasing lock "refresh_cache-107ed783-10a3-4f07-a809-a5a076c2b904" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.786343] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52819386-f9c8-df8c-8896-511a71627f88, 'name': SearchDatastore_Task, 'duration_secs': 0.016616} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.787664] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a0a2955-88ca-4354-9ea4-e71143f43b56 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.803397] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 892.803397] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ed2822-4b9d-5f40-d858-df386d4961cf" [ 892.803397] env[62600]: _type = "Task" [ 892.803397] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.812753] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ed2822-4b9d-5f40-d858-df386d4961cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.869929] env[62600]: DEBUG oslo_vmware.api [None req-5ddefda0-b3d6-4d04-afc4-3cf39aa5a1da tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222630, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.937755] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222625, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.003062] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ffb72a1-7d3f-4515-8dc0-18602ae476dc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.006234] env[62600]: DEBUG oslo_concurrency.lockutils [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "refresh_cache-d627a701-77cd-4a1d-9e52-0fc4adc93391" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.006421] env[62600]: DEBUG oslo_concurrency.lockutils [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "refresh_cache-d627a701-77cd-4a1d-9e52-0fc4adc93391" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.006590] env[62600]: DEBUG nova.network.neutron [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 893.014651] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8878daba-a0ff-43d5-a033-ce21cf11964d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.052403] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-003d66cd-50e1-46ed-bdfe-8d405591a203 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.064190] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8787b40b-ab64-4157-80b4-37f4e444b298 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.068247] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222628, 'name': Rename_Task, 'duration_secs': 1.25488} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.068906] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 893.069765] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f0def35-6c72-4489-ae36-eb50c28c56de {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.080203] env[62600]: DEBUG nova.compute.provider_tree [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.082848] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 893.082848] env[62600]: value = "task-1222632" [ 893.082848] env[62600]: _type = "Task" [ 893.082848] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.092360] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222632, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.115039] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': task-1222631, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089065} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.115380] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.116248] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-420d9534-a48d-4192-a5f6-e3d9c8800c2c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.140845] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] f6957b5a-5da8-4205-92e5-2f08bf948e88/f6957b5a-5da8-4205-92e5-2f08bf948e88.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.141201] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9885f425-ad0f-4380-b509-5a39c3a6c9c1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.162926] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Waiting for the task: (returnval){ [ 893.162926] env[62600]: value = "task-1222633" [ 893.162926] env[62600]: _type = "Task" [ 893.162926] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.176144] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': task-1222633, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.314086] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ed2822-4b9d-5f40-d858-df386d4961cf, 'name': SearchDatastore_Task, 'duration_secs': 0.017397} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.314438] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.314725] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 107ed783-10a3-4f07-a809-a5a076c2b904/107ed783-10a3-4f07-a809-a5a076c2b904.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 893.315030] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6ff1889a-17c0-4973-acf2-586bf715b411 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.324133] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 893.324133] env[62600]: value = "task-1222634" [ 893.324133] env[62600]: _type = "Task" [ 893.324133] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.333271] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222634, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.368523] env[62600]: DEBUG oslo_vmware.api [None req-5ddefda0-b3d6-4d04-afc4-3cf39aa5a1da tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222630, 'name': ReconfigVM_Task, 'duration_secs': 1.066615} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.368846] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ddefda0-b3d6-4d04-afc4-3cf39aa5a1da tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264299', 'volume_id': '93edc5a4-1bb4-44c9-868a-43a307466fe0', 'name': 'volume-93edc5a4-1bb4-44c9-868a-43a307466fe0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dde6df10-618a-40a8-b33f-efc0ca3a9287', 'attached_at': '', 'detached_at': '', 'volume_id': '93edc5a4-1bb4-44c9-868a-43a307466fe0', 'serial': '93edc5a4-1bb4-44c9-868a-43a307466fe0'} {{(pid=62600) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 893.438033] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222625, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.374291} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.438175] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e1395438-00bb-4e48-add8-ef3068f3c1c3/e1395438-00bb-4e48-add8-ef3068f3c1c3.vmdk to [datastore2] c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f/c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 893.438979] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-433c813d-41fa-4259-a6d9-abe4177c606b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.464362] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f/c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f.vmdk or device None with type streamOptimized {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.464929] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-86294cbd-7a74-4f0a-a685-5ae2ae57384f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.490454] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 893.490454] env[62600]: value = "task-1222635" [ 893.490454] env[62600]: _type = "Task" [ 893.490454] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.503408] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222635, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.559327] env[62600]: DEBUG nova.network.neutron [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 893.584322] env[62600]: DEBUG nova.scheduler.client.report [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.598998] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222632, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.680106] env[62600]: DEBUG nova.compute.manager [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 893.683317] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': task-1222633, 'name': ReconfigVM_Task, 'duration_secs': 0.362026} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.684342] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Reconfigured VM instance instance-00000047 to attach disk [datastore1] f6957b5a-5da8-4205-92e5-2f08bf948e88/f6957b5a-5da8-4205-92e5-2f08bf948e88.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 893.686674] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-37d90081-06a7-4b59-9777-da14e840f6de {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.697535] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Waiting for the task: (returnval){ [ 893.697535] env[62600]: value = "task-1222636" [ 893.697535] env[62600]: _type = "Task" [ 893.697535] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.713633] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': task-1222636, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.720815] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 893.722166] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 893.722166] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.722677] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 893.722905] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.723123] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 893.723528] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 893.724130] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 893.724130] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 893.724244] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 893.724512] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.726739] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b44acbe4-bc71-4699-b762-4b08bf2911a1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.737724] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-683e2936-7d13-4451-af90-21eb92ff4dff {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.822652] env[62600]: DEBUG nova.network.neutron [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Updating instance_info_cache with network_info: [{"id": "92bc4652-839c-47b3-bcf5-2cc0c43d5537", "address": "fa:16:3e:16:7c:4d", "network": {"id": "95818650-8227-4d11-9d14-007f06c07790", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1738152115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73e664d178f7484a9f4741b4d9450e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92bc4652-83", "ovs_interfaceid": "92bc4652-839c-47b3-bcf5-2cc0c43d5537", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.838425] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222634, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.930967] env[62600]: DEBUG nova.objects.instance [None req-5ddefda0-b3d6-4d04-afc4-3cf39aa5a1da tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lazy-loading 'flavor' on Instance uuid dde6df10-618a-40a8-b33f-efc0ca3a9287 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.985102] env[62600]: DEBUG nova.compute.manager [req-549cb95e-6e48-4720-9463-72d56468bf9d req-a1ba9605-e0d6-48d1-bf37-18d02cf42bd2 service nova] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Received event network-vif-plugged-e04d7519-77e4-4593-b7ec-5fc549cfce8f {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 893.985102] env[62600]: DEBUG oslo_concurrency.lockutils [req-549cb95e-6e48-4720-9463-72d56468bf9d req-a1ba9605-e0d6-48d1-bf37-18d02cf42bd2 service nova] Acquiring lock "d2e43ba2-4ccf-4625-91ba-78a6f2632461-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.985102] env[62600]: DEBUG oslo_concurrency.lockutils [req-549cb95e-6e48-4720-9463-72d56468bf9d req-a1ba9605-e0d6-48d1-bf37-18d02cf42bd2 service nova] Lock "d2e43ba2-4ccf-4625-91ba-78a6f2632461-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.985102] env[62600]: DEBUG oslo_concurrency.lockutils [req-549cb95e-6e48-4720-9463-72d56468bf9d req-a1ba9605-e0d6-48d1-bf37-18d02cf42bd2 service nova] Lock "d2e43ba2-4ccf-4625-91ba-78a6f2632461-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.985102] env[62600]: DEBUG nova.compute.manager [req-549cb95e-6e48-4720-9463-72d56468bf9d req-a1ba9605-e0d6-48d1-bf37-18d02cf42bd2 service nova] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] No waiting events found dispatching network-vif-plugged-e04d7519-77e4-4593-b7ec-5fc549cfce8f {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 893.985102] env[62600]: WARNING nova.compute.manager [req-549cb95e-6e48-4720-9463-72d56468bf9d req-a1ba9605-e0d6-48d1-bf37-18d02cf42bd2 service nova] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Received unexpected event network-vif-plugged-e04d7519-77e4-4593-b7ec-5fc549cfce8f for instance with vm_state building and task_state spawning. [ 894.001972] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222635, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.066248] env[62600]: DEBUG nova.network.neutron [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Successfully updated port: e04d7519-77e4-4593-b7ec-5fc549cfce8f {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 894.092741] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.093543] env[62600]: DEBUG nova.compute.manager [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 894.781836] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.792s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.782133] env[62600]: DEBUG nova.objects.instance [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lazy-loading 'resources' on Instance uuid efff4d72-4673-4a8d-9a81-be9d700ff881 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.783536] env[62600]: DEBUG oslo_concurrency.lockutils [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "refresh_cache-d627a701-77cd-4a1d-9e52-0fc4adc93391" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.783794] env[62600]: DEBUG nova.compute.manager [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Instance network_info: |[{"id": "92bc4652-839c-47b3-bcf5-2cc0c43d5537", "address": "fa:16:3e:16:7c:4d", "network": {"id": "95818650-8227-4d11-9d14-007f06c07790", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1738152115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73e664d178f7484a9f4741b4d9450e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92bc4652-83", "ovs_interfaceid": "92bc4652-839c-47b3-bcf5-2cc0c43d5537", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 894.785870] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "refresh_cache-d2e43ba2-4ccf-4625-91ba-78a6f2632461" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.786011] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquired lock "refresh_cache-d2e43ba2-4ccf-4625-91ba-78a6f2632461" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.786158] env[62600]: DEBUG nova.network.neutron [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 894.788162] env[62600]: DEBUG nova.compute.manager [req-0b6e98fe-a36b-4cad-b702-dc4d58421381 req-c98e964e-b0fa-4eb4-83a5-f0a3877421dd service nova] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Received event network-changed-92bc4652-839c-47b3-bcf5-2cc0c43d5537 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 894.788351] env[62600]: DEBUG nova.compute.manager [req-0b6e98fe-a36b-4cad-b702-dc4d58421381 req-c98e964e-b0fa-4eb4-83a5-f0a3877421dd service nova] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Refreshing instance network info cache due to event network-changed-92bc4652-839c-47b3-bcf5-2cc0c43d5537. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 894.788555] env[62600]: DEBUG oslo_concurrency.lockutils [req-0b6e98fe-a36b-4cad-b702-dc4d58421381 req-c98e964e-b0fa-4eb4-83a5-f0a3877421dd service nova] Acquiring lock "refresh_cache-d627a701-77cd-4a1d-9e52-0fc4adc93391" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.788696] env[62600]: DEBUG oslo_concurrency.lockutils [req-0b6e98fe-a36b-4cad-b702-dc4d58421381 req-c98e964e-b0fa-4eb4-83a5-f0a3877421dd service nova] Acquired lock "refresh_cache-d627a701-77cd-4a1d-9e52-0fc4adc93391" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.788853] env[62600]: DEBUG nova.network.neutron [req-0b6e98fe-a36b-4cad-b702-dc4d58421381 req-c98e964e-b0fa-4eb4-83a5-f0a3877421dd service nova] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Refreshing network info cache for port 92bc4652-839c-47b3-bcf5-2cc0c43d5537 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 894.802262] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:16:7c:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43798f54-0c07-4417-a23f-58bb6b7e204b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '92bc4652-839c-47b3-bcf5-2cc0c43d5537', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 894.809925] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Creating folder: Project (73e664d178f7484a9f4741b4d9450e68). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 894.815828] env[62600]: DEBUG oslo_vmware.api [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222632, 'name': PowerOnVM_Task, 'duration_secs': 0.738459} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.817799] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5ddefda0-b3d6-4d04-afc4-3cf39aa5a1da tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.215s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.818990] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a64ead4d-bdb7-428f-a1c7-109503326f1a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.820483] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 894.820686] env[62600]: INFO nova.compute.manager [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Took 11.00 seconds to spawn the instance on the hypervisor. [ 894.820867] env[62600]: DEBUG nova.compute.manager [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 894.823341] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b58fea6f-7e24-4ad3-9711-ab03e985e772 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.836438] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222634, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.531619} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.836855] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222635, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.837242] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': task-1222636, 'name': Rename_Task, 'duration_secs': 0.281458} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.838656] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 107ed783-10a3-4f07-a809-a5a076c2b904/107ed783-10a3-4f07-a809-a5a076c2b904.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 894.839035] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 894.839758] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 894.843987] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1c05e15c-a272-4135-ab98-9042f53162f6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.847431] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-af19f383-80b7-44d6-993b-95975c2358c2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.855587] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Created folder: Project (73e664d178f7484a9f4741b4d9450e68) in parent group-v264198. [ 894.855780] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Creating folder: Instances. Parent ref: group-v264313. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 894.857407] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-44185d7d-89dc-4b60-ad8e-a84b84a8010e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.863016] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Waiting for the task: (returnval){ [ 894.863016] env[62600]: value = "task-1222639" [ 894.863016] env[62600]: _type = "Task" [ 894.863016] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.869020] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 894.869020] env[62600]: value = "task-1222638" [ 894.869020] env[62600]: _type = "Task" [ 894.869020] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.870810] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Created folder: Instances in parent group-v264313. [ 894.871071] env[62600]: DEBUG oslo.service.loopingcall [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 894.874792] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 894.878657] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-97fd52cb-94eb-405e-954d-9f61a2232750 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.897588] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': task-1222639, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.911209] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222638, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.912842] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 894.912842] env[62600]: value = "task-1222641" [ 894.912842] env[62600]: _type = "Task" [ 894.912842] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.926227] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222641, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.149021] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8aed804-8579-4090-ba16-a61582bb0b4f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.154351] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4626002c-87ec-40e9-a8dc-587ada8f242d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.185060] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b72e9416-2474-479b-8a16-eb1819274021 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.193491] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d9038e5-a1ca-4deb-85fe-1711bb4f3a11 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.207288] env[62600]: DEBUG nova.compute.provider_tree [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.297599] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222635, 'name': ReconfigVM_Task, 'duration_secs': 1.478338} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.297851] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Reconfigured VM instance instance-00000045 to attach disk [datastore2] c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f/c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f.vmdk or device None with type streamOptimized {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 895.299486] env[62600]: DEBUG nova.compute.utils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 895.300758] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa3c1369-3e88-42ee-9fb6-2d52e832ac56 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.302695] env[62600]: DEBUG nova.compute.manager [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 895.302864] env[62600]: DEBUG nova.network.neutron [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 895.312677] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 895.312677] env[62600]: value = "task-1222642" [ 895.312677] env[62600]: _type = "Task" [ 895.312677] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.329085] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222642, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.336203] env[62600]: DEBUG nova.network.neutron [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 895.365085] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8fb59274-8c0f-4f6b-bff3-8cfc25390b0e tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquiring lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.365804] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8fb59274-8c0f-4f6b-bff3-8cfc25390b0e tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.381777] env[62600]: INFO nova.compute.manager [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Took 33.40 seconds to build instance. [ 895.392055] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': task-1222639, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.397178] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222638, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074627} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.397974] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 895.398471] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e4f0361-04c0-46e3-82bc-f87f186a06bb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.404391] env[62600]: DEBUG nova.policy [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '41f578fc709a497abd53d1a4f0540f6d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5e7eca5e197e47759eeea423e280986f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 895.429554] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] 107ed783-10a3-4f07-a809-a5a076c2b904/107ed783-10a3-4f07-a809-a5a076c2b904.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 895.437600] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1be197a4-09d4-453e-9cbd-550e463071d2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.460343] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222641, 'name': CreateVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.461898] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 895.461898] env[62600]: value = "task-1222643" [ 895.461898] env[62600]: _type = "Task" [ 895.461898] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.472727] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222643, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.648674] env[62600]: DEBUG nova.network.neutron [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Updating instance_info_cache with network_info: [{"id": "e04d7519-77e4-4593-b7ec-5fc549cfce8f", "address": "fa:16:3e:4f:f4:85", "network": {"id": "353f7a99-4652-4a0d-be26-539c5c42b3d2", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1505004576-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7eca5e197e47759eeea423e280986f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape04d7519-77", "ovs_interfaceid": "e04d7519-77e4-4593-b7ec-5fc549cfce8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.710714] env[62600]: DEBUG nova.scheduler.client.report [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.732831] env[62600]: DEBUG nova.network.neutron [req-0b6e98fe-a36b-4cad-b702-dc4d58421381 req-c98e964e-b0fa-4eb4-83a5-f0a3877421dd service nova] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Updated VIF entry in instance network info cache for port 92bc4652-839c-47b3-bcf5-2cc0c43d5537. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 895.734084] env[62600]: DEBUG nova.network.neutron [req-0b6e98fe-a36b-4cad-b702-dc4d58421381 req-c98e964e-b0fa-4eb4-83a5-f0a3877421dd service nova] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Updating instance_info_cache with network_info: [{"id": "92bc4652-839c-47b3-bcf5-2cc0c43d5537", "address": "fa:16:3e:16:7c:4d", "network": {"id": "95818650-8227-4d11-9d14-007f06c07790", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1738152115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73e664d178f7484a9f4741b4d9450e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92bc4652-83", "ovs_interfaceid": "92bc4652-839c-47b3-bcf5-2cc0c43d5537", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.806742] env[62600]: DEBUG nova.compute.manager [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 895.823993] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222642, 'name': Rename_Task, 'duration_secs': 0.211309} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.823993] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 895.824240] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f9cd3a71-2852-47b1-ac5a-437bc55b4ca2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.831172] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 895.831172] env[62600]: value = "task-1222644" [ 895.831172] env[62600]: _type = "Task" [ 895.831172] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.839420] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222644, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.874338] env[62600]: INFO nova.compute.manager [None req-8fb59274-8c0f-4f6b-bff3-8cfc25390b0e tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Detaching volume 83180b95-0343-4a55-9409-831dfcf80c02 [ 895.882945] env[62600]: DEBUG oslo_vmware.api [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': task-1222639, 'name': PowerOnVM_Task, 'duration_secs': 0.532203} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.885727] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 895.885953] env[62600]: INFO nova.compute.manager [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Took 9.58 seconds to spawn the instance on the hypervisor. [ 895.886165] env[62600]: DEBUG nova.compute.manager [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.886592] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fe7f45e9-a81f-4e5a-857c-cac118865823 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "ab11cdd1-80a7-43c4-afac-24c7930238c1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.914s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.887935] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3261ab83-14a0-42e0-85c8-46bf3942bbf0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.932097] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222641, 'name': CreateVM_Task, 'duration_secs': 0.62109} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.933620] env[62600]: INFO nova.virt.block_device [None req-8fb59274-8c0f-4f6b-bff3-8cfc25390b0e tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Attempting to driver detach volume 83180b95-0343-4a55-9409-831dfcf80c02 from mountpoint /dev/sdc [ 895.933913] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fb59274-8c0f-4f6b-bff3-8cfc25390b0e tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Volume detach. Driver type: vmdk {{(pid=62600) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 895.934076] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fb59274-8c0f-4f6b-bff3-8cfc25390b0e tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264303', 'volume_id': '83180b95-0343-4a55-9409-831dfcf80c02', 'name': 'volume-83180b95-0343-4a55-9409-831dfcf80c02', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dde6df10-618a-40a8-b33f-efc0ca3a9287', 'attached_at': '', 'detached_at': '', 'volume_id': '83180b95-0343-4a55-9409-831dfcf80c02', 'serial': '83180b95-0343-4a55-9409-831dfcf80c02'} {{(pid=62600) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 895.934453] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 895.935843] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53782e9d-74ac-43c4-871a-77023f416c07 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.939320] env[62600]: DEBUG oslo_concurrency.lockutils [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.939489] env[62600]: DEBUG oslo_concurrency.lockutils [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.939804] env[62600]: DEBUG oslo_concurrency.lockutils [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 895.940342] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c2c3968-cee3-41d8-a914-b95d5d8b614e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.946035] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 895.946035] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52963c80-a1b0-d6c8-da50-0d4f44696623" [ 895.946035] env[62600]: _type = "Task" [ 895.946035] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.971321] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce5a27c2-b6b4-4e4a-bba4-03dc4a27c032 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.979863] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222643, 'name': ReconfigVM_Task, 'duration_secs': 0.367929} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.984864] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Reconfigured VM instance instance-00000048 to attach disk [datastore1] 107ed783-10a3-4f07-a809-a5a076c2b904/107ed783-10a3-4f07-a809-a5a076c2b904.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 895.985569] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52963c80-a1b0-d6c8-da50-0d4f44696623, 'name': SearchDatastore_Task, 'duration_secs': 0.020441} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.985925] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ef12cf19-c3c0-425f-ba5e-cf8382f51a41 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.987564] env[62600]: DEBUG oslo_concurrency.lockutils [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.988126] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 895.988389] env[62600]: DEBUG oslo_concurrency.lockutils [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.988548] env[62600]: DEBUG oslo_concurrency.lockutils [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.988734] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 895.989547] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5aa49cc-fab0-4a1b-a38a-3f42cf3d157c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.991920] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8965466b-a478-48a7-addc-2e9a026cd3f3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.016682] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aee377bb-ee06-483d-8913-2ab80cdae688 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.019422] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 896.019638] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 896.020583] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 896.020583] env[62600]: value = "task-1222645" [ 896.020583] env[62600]: _type = "Task" [ 896.020583] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.020807] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce5b77a6-e53a-4128-991b-c45cf260ca4e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.039106] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fb59274-8c0f-4f6b-bff3-8cfc25390b0e tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] The volume has not been displaced from its original location: [datastore2] volume-83180b95-0343-4a55-9409-831dfcf80c02/volume-83180b95-0343-4a55-9409-831dfcf80c02.vmdk. No consolidation needed. {{(pid=62600) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 896.044856] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fb59274-8c0f-4f6b-bff3-8cfc25390b0e tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Reconfiguring VM instance instance-0000003b to detach disk 2002 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 896.046917] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aa5c717b-2ce9-412a-ab7d-34df7fa2fb3a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.061488] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 896.061488] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d710c1-620d-9f0e-73e2-856825122793" [ 896.061488] env[62600]: _type = "Task" [ 896.061488] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.065576] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222645, 'name': Rename_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.071806] env[62600]: DEBUG oslo_vmware.api [None req-8fb59274-8c0f-4f6b-bff3-8cfc25390b0e tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for the task: (returnval){ [ 896.071806] env[62600]: value = "task-1222646" [ 896.071806] env[62600]: _type = "Task" [ 896.071806] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.079304] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d710c1-620d-9f0e-73e2-856825122793, 'name': SearchDatastore_Task, 'duration_secs': 0.010471} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.080625] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c1a7367-3fe1-4bbe-89eb-6cfa01617aa9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.087213] env[62600]: DEBUG oslo_vmware.api [None req-8fb59274-8c0f-4f6b-bff3-8cfc25390b0e tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222646, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.091100] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 896.091100] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]529ed55e-4287-a68c-d8d6-c9ede4f14458" [ 896.091100] env[62600]: _type = "Task" [ 896.091100] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.100699] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]529ed55e-4287-a68c-d8d6-c9ede4f14458, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.151480] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Releasing lock "refresh_cache-d2e43ba2-4ccf-4625-91ba-78a6f2632461" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.151884] env[62600]: DEBUG nova.compute.manager [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Instance network_info: |[{"id": "e04d7519-77e4-4593-b7ec-5fc549cfce8f", "address": "fa:16:3e:4f:f4:85", "network": {"id": "353f7a99-4652-4a0d-be26-539c5c42b3d2", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1505004576-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7eca5e197e47759eeea423e280986f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape04d7519-77", "ovs_interfaceid": "e04d7519-77e4-4593-b7ec-5fc549cfce8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 896.152404] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:f4:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd998416-f3d6-4a62-b828-5011063ce76a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e04d7519-77e4-4593-b7ec-5fc549cfce8f', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 896.160619] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Creating folder: Project (5e7eca5e197e47759eeea423e280986f). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 896.160978] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3ccae99f-f75d-42f6-9311-57c7b7309faa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.176439] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Created folder: Project (5e7eca5e197e47759eeea423e280986f) in parent group-v264198. [ 896.176657] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Creating folder: Instances. Parent ref: group-v264316. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 896.176909] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-095f2d1c-7a99-4313-b97d-e4e66f9e7421 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.190072] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Created folder: Instances in parent group-v264316. [ 896.190470] env[62600]: DEBUG oslo.service.loopingcall [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 896.190712] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 896.190912] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6cd0e55a-7aee-48fe-bd46-212e399084c4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.207630] env[62600]: DEBUG nova.network.neutron [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Successfully created port: f01caa0e-4aff-41e9-a7ed-0511436b131d {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 896.216213] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.434s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.219726] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.109s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.221329] env[62600]: INFO nova.compute.claims [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 896.226072] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 896.226072] env[62600]: value = "task-1222649" [ 896.226072] env[62600]: _type = "Task" [ 896.226072] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.238027] env[62600]: DEBUG oslo_concurrency.lockutils [req-0b6e98fe-a36b-4cad-b702-dc4d58421381 req-c98e964e-b0fa-4eb4-83a5-f0a3877421dd service nova] Releasing lock "refresh_cache-d627a701-77cd-4a1d-9e52-0fc4adc93391" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.238027] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222649, 'name': CreateVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.247435] env[62600]: INFO nova.scheduler.client.report [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Deleted allocations for instance efff4d72-4673-4a8d-9a81-be9d700ff881 [ 896.350957] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222644, 'name': PowerOnVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.366815] env[62600]: DEBUG nova.compute.manager [req-1956fb58-6538-4282-b5bd-4eeebd8c792a req-8326977d-c71f-4c32-9d1d-5ac59fe55c41 service nova] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Received event network-changed-e04d7519-77e4-4593-b7ec-5fc549cfce8f {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.368435] env[62600]: DEBUG nova.compute.manager [req-1956fb58-6538-4282-b5bd-4eeebd8c792a req-8326977d-c71f-4c32-9d1d-5ac59fe55c41 service nova] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Refreshing instance network info cache due to event network-changed-e04d7519-77e4-4593-b7ec-5fc549cfce8f. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 896.368435] env[62600]: DEBUG oslo_concurrency.lockutils [req-1956fb58-6538-4282-b5bd-4eeebd8c792a req-8326977d-c71f-4c32-9d1d-5ac59fe55c41 service nova] Acquiring lock "refresh_cache-d2e43ba2-4ccf-4625-91ba-78a6f2632461" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.368435] env[62600]: DEBUG oslo_concurrency.lockutils [req-1956fb58-6538-4282-b5bd-4eeebd8c792a req-8326977d-c71f-4c32-9d1d-5ac59fe55c41 service nova] Acquired lock "refresh_cache-d2e43ba2-4ccf-4625-91ba-78a6f2632461" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.368435] env[62600]: DEBUG nova.network.neutron [req-1956fb58-6538-4282-b5bd-4eeebd8c792a req-8326977d-c71f-4c32-9d1d-5ac59fe55c41 service nova] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Refreshing network info cache for port e04d7519-77e4-4593-b7ec-5fc549cfce8f {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 896.396016] env[62600]: DEBUG nova.compute.manager [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 896.411964] env[62600]: INFO nova.compute.manager [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Took 33.12 seconds to build instance. [ 896.462432] env[62600]: DEBUG nova.compute.manager [req-f211d651-7e16-4759-8816-06863214117d req-95fc59df-aa4a-4e1b-8890-3ce44a1e9c36 service nova] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Received event network-changed-517fe831-7f4c-477a-9c43-496ec9895bb3 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.462432] env[62600]: DEBUG nova.compute.manager [req-f211d651-7e16-4759-8816-06863214117d req-95fc59df-aa4a-4e1b-8890-3ce44a1e9c36 service nova] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Refreshing instance network info cache due to event network-changed-517fe831-7f4c-477a-9c43-496ec9895bb3. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 896.462432] env[62600]: DEBUG oslo_concurrency.lockutils [req-f211d651-7e16-4759-8816-06863214117d req-95fc59df-aa4a-4e1b-8890-3ce44a1e9c36 service nova] Acquiring lock "refresh_cache-ab11cdd1-80a7-43c4-afac-24c7930238c1" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.462432] env[62600]: DEBUG oslo_concurrency.lockutils [req-f211d651-7e16-4759-8816-06863214117d req-95fc59df-aa4a-4e1b-8890-3ce44a1e9c36 service nova] Acquired lock "refresh_cache-ab11cdd1-80a7-43c4-afac-24c7930238c1" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.462432] env[62600]: DEBUG nova.network.neutron [req-f211d651-7e16-4759-8816-06863214117d req-95fc59df-aa4a-4e1b-8890-3ce44a1e9c36 service nova] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Refreshing network info cache for port 517fe831-7f4c-477a-9c43-496ec9895bb3 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 896.538110] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222645, 'name': Rename_Task, 'duration_secs': 0.164516} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.538531] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 896.538850] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f44af747-d9ec-4ad0-887d-d8557bf7a774 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.548509] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 896.548509] env[62600]: value = "task-1222650" [ 896.548509] env[62600]: _type = "Task" [ 896.548509] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.570434] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222650, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.589967] env[62600]: DEBUG oslo_vmware.api [None req-8fb59274-8c0f-4f6b-bff3-8cfc25390b0e tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222646, 'name': ReconfigVM_Task, 'duration_secs': 0.269239} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.590186] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fb59274-8c0f-4f6b-bff3-8cfc25390b0e tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Reconfigured VM instance instance-0000003b to detach disk 2002 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 896.596370] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fc556a32-e1d1-4446-ad20-7263f103c930 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.623307] env[62600]: DEBUG oslo_vmware.api [None req-8fb59274-8c0f-4f6b-bff3-8cfc25390b0e tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for the task: (returnval){ [ 896.623307] env[62600]: value = "task-1222651" [ 896.623307] env[62600]: _type = "Task" [ 896.623307] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.633617] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]529ed55e-4287-a68c-d8d6-c9ede4f14458, 'name': SearchDatastore_Task, 'duration_secs': 0.012902} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.634496] env[62600]: DEBUG oslo_concurrency.lockutils [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.635264] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] d627a701-77cd-4a1d-9e52-0fc4adc93391/d627a701-77cd-4a1d-9e52-0fc4adc93391.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 896.635403] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-34dd48cd-1fbf-4048-9cef-dd24f2e530b5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.641285] env[62600]: DEBUG oslo_vmware.api [None req-8fb59274-8c0f-4f6b-bff3-8cfc25390b0e tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222651, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.642989] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 896.642989] env[62600]: value = "task-1222652" [ 896.642989] env[62600]: _type = "Task" [ 896.642989] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.656980] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222652, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.740226] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222649, 'name': CreateVM_Task, 'duration_secs': 0.4918} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.741632] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 896.742022] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.742321] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.742736] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 896.743271] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f9f6c3e-73d3-48af-beea-3560760efb64 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.749069] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 896.749069] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]524582ab-77f4-4ca0-9359-a0f73d5cb743" [ 896.749069] env[62600]: _type = "Task" [ 896.749069] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.757395] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3c86ae36-ed0f-4550-b519-4a3a739ded7c tempest-ListImageFiltersTestJSON-90386634 tempest-ListImageFiltersTestJSON-90386634-project-member] Lock "efff4d72-4673-4a8d-9a81-be9d700ff881" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.860s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.772821] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]524582ab-77f4-4ca0-9359-a0f73d5cb743, 'name': SearchDatastore_Task, 'duration_secs': 0.018988} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.773953] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.774726] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 896.774726] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.774726] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.774896] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 896.775356] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-80aa794f-ba84-4ef5-ade0-a29847acaeeb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.785981] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 896.786212] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 896.787050] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcbd26e8-7490-411e-9106-0bc592c4a6af {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.796643] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 896.796643] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ccd8d7-250f-69ab-0d11-c0658ed32e25" [ 896.796643] env[62600]: _type = "Task" [ 896.796643] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.806098] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ccd8d7-250f-69ab-0d11-c0658ed32e25, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.819225] env[62600]: DEBUG nova.compute.manager [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 896.843224] env[62600]: DEBUG oslo_vmware.api [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222644, 'name': PowerOnVM_Task, 'duration_secs': 0.573402} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.845444] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 896.845675] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 896.845841] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 896.846038] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 896.846194] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 896.846504] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 896.846574] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 896.846712] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 896.846883] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 896.847062] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 896.847243] env[62600]: DEBUG nova.virt.hardware [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 896.847547] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 896.847790] env[62600]: INFO nova.compute.manager [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Took 15.48 seconds to spawn the instance on the hypervisor. [ 896.847928] env[62600]: DEBUG nova.compute.manager [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.848682] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8cab3ba-bb13-4927-bae2-cb06f8d4f32f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.851685] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95a73c92-84fb-4825-9db4-95bb93d06172 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.864858] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdca226e-c9d5-44a1-9db9-b0654bbfd69e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.917504] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7d0e3b7c-129a-477f-b697-74cc753c57b9 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Lock "f6957b5a-5da8-4205-92e5-2f08bf948e88" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.645s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.918471] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.059075] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222650, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.136337] env[62600]: DEBUG oslo_vmware.api [None req-8fb59274-8c0f-4f6b-bff3-8cfc25390b0e tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222651, 'name': ReconfigVM_Task, 'duration_secs': 0.199014} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.136337] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fb59274-8c0f-4f6b-bff3-8cfc25390b0e tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264303', 'volume_id': '83180b95-0343-4a55-9409-831dfcf80c02', 'name': 'volume-83180b95-0343-4a55-9409-831dfcf80c02', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'dde6df10-618a-40a8-b33f-efc0ca3a9287', 'attached_at': '', 'detached_at': '', 'volume_id': '83180b95-0343-4a55-9409-831dfcf80c02', 'serial': '83180b95-0343-4a55-9409-831dfcf80c02'} {{(pid=62600) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 897.157642] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222652, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.310604] env[62600]: DEBUG nova.network.neutron [req-1956fb58-6538-4282-b5bd-4eeebd8c792a req-8326977d-c71f-4c32-9d1d-5ac59fe55c41 service nova] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Updated VIF entry in instance network info cache for port e04d7519-77e4-4593-b7ec-5fc549cfce8f. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 897.311090] env[62600]: DEBUG nova.network.neutron [req-1956fb58-6538-4282-b5bd-4eeebd8c792a req-8326977d-c71f-4c32-9d1d-5ac59fe55c41 service nova] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Updating instance_info_cache with network_info: [{"id": "e04d7519-77e4-4593-b7ec-5fc549cfce8f", "address": "fa:16:3e:4f:f4:85", "network": {"id": "353f7a99-4652-4a0d-be26-539c5c42b3d2", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1505004576-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7eca5e197e47759eeea423e280986f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape04d7519-77", "ovs_interfaceid": "e04d7519-77e4-4593-b7ec-5fc549cfce8f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.324342] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ccd8d7-250f-69ab-0d11-c0658ed32e25, 'name': SearchDatastore_Task, 'duration_secs': 0.015581} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.325826] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48b813e6-a068-4b44-8dcd-aadae951b72a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.333454] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 897.333454] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5201b2e0-a00b-3bc4-bfe7-a86c5149d5c1" [ 897.333454] env[62600]: _type = "Task" [ 897.333454] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.345531] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5201b2e0-a00b-3bc4-bfe7-a86c5149d5c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.371931] env[62600]: INFO nova.compute.manager [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Took 36.68 seconds to build instance. [ 897.397856] env[62600]: DEBUG nova.network.neutron [req-f211d651-7e16-4759-8816-06863214117d req-95fc59df-aa4a-4e1b-8890-3ce44a1e9c36 service nova] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Updated VIF entry in instance network info cache for port 517fe831-7f4c-477a-9c43-496ec9895bb3. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 897.398315] env[62600]: DEBUG nova.network.neutron [req-f211d651-7e16-4759-8816-06863214117d req-95fc59df-aa4a-4e1b-8890-3ce44a1e9c36 service nova] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Updating instance_info_cache with network_info: [{"id": "517fe831-7f4c-477a-9c43-496ec9895bb3", "address": "fa:16:3e:74:95:a0", "network": {"id": "e7547ac4-24a6-4778-a17a-40e79e7a0d0c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-320671681-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3bb131f3994f7a8673dafff6d353c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap517fe831-7f", "ovs_interfaceid": "517fe831-7f4c-477a-9c43-496ec9895bb3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.420501] env[62600]: DEBUG nova.compute.manager [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 897.562249] env[62600]: DEBUG oslo_vmware.api [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222650, 'name': PowerOnVM_Task, 'duration_secs': 0.934169} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.562587] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 897.562803] env[62600]: INFO nova.compute.manager [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Took 8.76 seconds to spawn the instance on the hypervisor. [ 897.562990] env[62600]: DEBUG nova.compute.manager [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 897.563855] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd574c9-bfbe-43e2-99bb-df45f00c329c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.592941] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5fa47aa-993a-4b96-9ba2-f8359c130e02 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.605077] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4224925a-928f-4d92-9cfa-312d086058e9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.645233] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e246ef42-fe9a-4548-8088-830268632de3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.657554] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222652, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.620258} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.659822] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] d627a701-77cd-4a1d-9e52-0fc4adc93391/d627a701-77cd-4a1d-9e52-0fc4adc93391.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 897.660055] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 897.660661] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c38a873e-1917-41bd-9582-1fbbb5820815 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.663775] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ef9af68-07fb-4fbf-9524-2bdae78b26be {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.687274] env[62600]: DEBUG nova.compute.provider_tree [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 897.690863] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 897.690863] env[62600]: value = "task-1222653" [ 897.690863] env[62600]: _type = "Task" [ 897.690863] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.703294] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222653, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.721193] env[62600]: DEBUG nova.objects.instance [None req-8fb59274-8c0f-4f6b-bff3-8cfc25390b0e tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lazy-loading 'flavor' on Instance uuid dde6df10-618a-40a8-b33f-efc0ca3a9287 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.790875] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Acquiring lock "f6957b5a-5da8-4205-92e5-2f08bf948e88" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.790875] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Lock "f6957b5a-5da8-4205-92e5-2f08bf948e88" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.790875] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Acquiring lock "f6957b5a-5da8-4205-92e5-2f08bf948e88-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.790875] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Lock "f6957b5a-5da8-4205-92e5-2f08bf948e88-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.790875] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Lock "f6957b5a-5da8-4205-92e5-2f08bf948e88-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.791818] env[62600]: INFO nova.compute.manager [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Terminating instance [ 897.794037] env[62600]: DEBUG nova.compute.manager [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 897.794465] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 897.795423] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61d2b963-fc41-4e43-8f06-850b171a3e52 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.804903] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 897.805741] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-277a7714-71be-4b08-8950-25b91cb22780 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.813616] env[62600]: DEBUG oslo_vmware.api [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Waiting for the task: (returnval){ [ 897.813616] env[62600]: value = "task-1222654" [ 897.813616] env[62600]: _type = "Task" [ 897.813616] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.817288] env[62600]: DEBUG oslo_concurrency.lockutils [req-1956fb58-6538-4282-b5bd-4eeebd8c792a req-8326977d-c71f-4c32-9d1d-5ac59fe55c41 service nova] Releasing lock "refresh_cache-d2e43ba2-4ccf-4625-91ba-78a6f2632461" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.823708] env[62600]: DEBUG oslo_vmware.api [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': task-1222654, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.844017] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5201b2e0-a00b-3bc4-bfe7-a86c5149d5c1, 'name': SearchDatastore_Task, 'duration_secs': 0.013156} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.847273] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.847273] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] d2e43ba2-4ccf-4625-91ba-78a6f2632461/d2e43ba2-4ccf-4625-91ba-78a6f2632461.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 897.847273] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3cf708c4-39bb-4ad5-99f7-60858c0cd531 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.857026] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 897.857026] env[62600]: value = "task-1222655" [ 897.857026] env[62600]: _type = "Task" [ 897.857026] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.864535] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222655, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.874023] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3afaf793-69e1-49d7-b643-ca203091ed8e tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.206s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.901229] env[62600]: DEBUG oslo_concurrency.lockutils [req-f211d651-7e16-4759-8816-06863214117d req-95fc59df-aa4a-4e1b-8890-3ce44a1e9c36 service nova] Releasing lock "refresh_cache-ab11cdd1-80a7-43c4-afac-24c7930238c1" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.954864] env[62600]: DEBUG oslo_concurrency.lockutils [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.084883] env[62600]: INFO nova.compute.manager [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Took 32.72 seconds to build instance. [ 898.093183] env[62600]: DEBUG nova.compute.manager [req-0dd7911f-3059-4d7d-a7c6-f0906519baf0 req-911bf494-64e9-410f-af28-f0ce5b52a9bf service nova] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Received event network-vif-plugged-f01caa0e-4aff-41e9-a7ed-0511436b131d {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.093728] env[62600]: DEBUG oslo_concurrency.lockutils [req-0dd7911f-3059-4d7d-a7c6-f0906519baf0 req-911bf494-64e9-410f-af28-f0ce5b52a9bf service nova] Acquiring lock "48bbbdaf-496d-4780-b467-6b393257535d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.094251] env[62600]: DEBUG oslo_concurrency.lockutils [req-0dd7911f-3059-4d7d-a7c6-f0906519baf0 req-911bf494-64e9-410f-af28-f0ce5b52a9bf service nova] Lock "48bbbdaf-496d-4780-b467-6b393257535d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.094564] env[62600]: DEBUG oslo_concurrency.lockutils [req-0dd7911f-3059-4d7d-a7c6-f0906519baf0 req-911bf494-64e9-410f-af28-f0ce5b52a9bf service nova] Lock "48bbbdaf-496d-4780-b467-6b393257535d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.094874] env[62600]: DEBUG nova.compute.manager [req-0dd7911f-3059-4d7d-a7c6-f0906519baf0 req-911bf494-64e9-410f-af28-f0ce5b52a9bf service nova] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] No waiting events found dispatching network-vif-plugged-f01caa0e-4aff-41e9-a7ed-0511436b131d {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 898.095985] env[62600]: WARNING nova.compute.manager [req-0dd7911f-3059-4d7d-a7c6-f0906519baf0 req-911bf494-64e9-410f-af28-f0ce5b52a9bf service nova] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Received unexpected event network-vif-plugged-f01caa0e-4aff-41e9-a7ed-0511436b131d for instance with vm_state building and task_state spawning. [ 898.205180] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222653, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.194756} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.205496] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 898.206913] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9313d0c-c7aa-42ad-bcaf-38e0f89fd2b5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.211926] env[62600]: ERROR nova.scheduler.client.report [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [req-4c415de6-55c7-4717-af17-2d7d4b55f843] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 664af347-7147-4bf5-9019-9ae15cb4aa82. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-4c415de6-55c7-4717-af17-2d7d4b55f843"}]} [ 898.237075] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] d627a701-77cd-4a1d-9e52-0fc4adc93391/d627a701-77cd-4a1d-9e52-0fc4adc93391.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 898.239755] env[62600]: DEBUG nova.scheduler.client.report [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Refreshing inventories for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 898.242532] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c3ff185-805b-4f28-9ed1-72b516ae0636 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.270028] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 898.270028] env[62600]: value = "task-1222656" [ 898.270028] env[62600]: _type = "Task" [ 898.270028] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.278071] env[62600]: DEBUG nova.scheduler.client.report [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Updating ProviderTree inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 898.278071] env[62600]: DEBUG nova.compute.provider_tree [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 898.283071] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222656, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.293391] env[62600]: DEBUG nova.scheduler.client.report [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Refreshing aggregate associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, aggregates: None {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 898.314163] env[62600]: DEBUG nova.scheduler.client.report [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Refreshing trait associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 898.331671] env[62600]: DEBUG oslo_vmware.api [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': task-1222654, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.369086] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222655, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.549704] env[62600]: DEBUG nova.network.neutron [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Successfully updated port: f01caa0e-4aff-41e9-a7ed-0511436b131d {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 898.573427] env[62600]: DEBUG nova.compute.manager [req-b538ba70-f4f8-42c4-95de-ba046a948dad req-d8f7bb4b-dcf8-46ce-939e-874890bc5097 service nova] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Received event network-changed-f01caa0e-4aff-41e9-a7ed-0511436b131d {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.573598] env[62600]: DEBUG nova.compute.manager [req-b538ba70-f4f8-42c4-95de-ba046a948dad req-d8f7bb4b-dcf8-46ce-939e-874890bc5097 service nova] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Refreshing instance network info cache due to event network-changed-f01caa0e-4aff-41e9-a7ed-0511436b131d. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 898.573811] env[62600]: DEBUG oslo_concurrency.lockutils [req-b538ba70-f4f8-42c4-95de-ba046a948dad req-d8f7bb4b-dcf8-46ce-939e-874890bc5097 service nova] Acquiring lock "refresh_cache-48bbbdaf-496d-4780-b467-6b393257535d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.573980] env[62600]: DEBUG oslo_concurrency.lockutils [req-b538ba70-f4f8-42c4-95de-ba046a948dad req-d8f7bb4b-dcf8-46ce-939e-874890bc5097 service nova] Acquired lock "refresh_cache-48bbbdaf-496d-4780-b467-6b393257535d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.574177] env[62600]: DEBUG nova.network.neutron [req-b538ba70-f4f8-42c4-95de-ba046a948dad req-d8f7bb4b-dcf8-46ce-939e-874890bc5097 service nova] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Refreshing network info cache for port f01caa0e-4aff-41e9-a7ed-0511436b131d {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 898.588148] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fed2f632-b426-46ab-a18d-0cb0a7d83801 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "107ed783-10a3-4f07-a809-a5a076c2b904" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.236s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.642796] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf229538-cc43-46be-bcca-3fb28d23c18a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.655154] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f50cae7-d430-49f4-88cd-e44ba4cece86 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.660552] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.660866] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.661195] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.661408] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.661687] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.697302] env[62600]: INFO nova.compute.manager [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Terminating instance [ 898.699729] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dd3bb37-c66e-4c3e-ba17-b452eba07846 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.703048] env[62600]: DEBUG nova.compute.manager [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 898.703271] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 898.704330] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b1e8a8b-cc36-45ed-8a76-f7911ffd74f3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.714185] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 898.717035] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fb70d73e-43a9-4f65-b280-61053b282bd4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.720112] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-795e266c-a99d-4937-8c50-bae13c8bfd57 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.738792] env[62600]: DEBUG nova.compute.provider_tree [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 898.742342] env[62600]: DEBUG oslo_vmware.api [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 898.742342] env[62600]: value = "task-1222657" [ 898.742342] env[62600]: _type = "Task" [ 898.742342] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.753811] env[62600]: DEBUG oslo_vmware.api [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222657, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.760057] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8fb59274-8c0f-4f6b-bff3-8cfc25390b0e tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.394s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.779670] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222656, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.828139] env[62600]: DEBUG oslo_vmware.api [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': task-1222654, 'name': PowerOffVM_Task, 'duration_secs': 0.583445} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.828894] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 898.828894] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 898.828894] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c3eadba1-d4e5-47fd-a4b4-8aadad1e1df0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.868115] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222655, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.012439} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.868549] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] d2e43ba2-4ccf-4625-91ba-78a6f2632461/d2e43ba2-4ccf-4625-91ba-78a6f2632461.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 898.868853] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 898.869152] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e2433e9b-be44-4838-91e7-f0d06ef25467 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.878266] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 898.878266] env[62600]: value = "task-1222659" [ 898.878266] env[62600]: _type = "Task" [ 898.878266] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.887578] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222659, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.912863] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 898.912863] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 898.913110] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Deleting the datastore file [datastore1] f6957b5a-5da8-4205-92e5-2f08bf948e88 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 898.913239] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cb4e4960-ccdf-49d4-b675-2e9ad06d1bf9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.921352] env[62600]: DEBUG oslo_vmware.api [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Waiting for the task: (returnval){ [ 898.921352] env[62600]: value = "task-1222660" [ 898.921352] env[62600]: _type = "Task" [ 898.921352] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.930712] env[62600]: DEBUG oslo_vmware.api [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': task-1222660, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.055271] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "refresh_cache-48bbbdaf-496d-4780-b467-6b393257535d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.144959] env[62600]: DEBUG nova.network.neutron [req-b538ba70-f4f8-42c4-95de-ba046a948dad req-d8f7bb4b-dcf8-46ce-939e-874890bc5097 service nova] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 899.259894] env[62600]: DEBUG oslo_vmware.api [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222657, 'name': PowerOffVM_Task, 'duration_secs': 0.304263} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.260365] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 899.260553] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 899.260832] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0b734040-f10a-4159-bcbd-a3bcd98f871e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.282564] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222656, 'name': ReconfigVM_Task, 'duration_secs': 0.559266} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.282564] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Reconfigured VM instance instance-00000049 to attach disk [datastore2] d627a701-77cd-4a1d-9e52-0fc4adc93391/d627a701-77cd-4a1d-9e52-0fc4adc93391.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 899.283635] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-85b9a4e7-b51d-4f7a-af94-ed3c8e4de375 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.293350] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 899.293350] env[62600]: value = "task-1222662" [ 899.293350] env[62600]: _type = "Task" [ 899.293350] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.297916] env[62600]: DEBUG nova.scheduler.client.report [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 99 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 899.298214] env[62600]: DEBUG nova.compute.provider_tree [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 99 to 100 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 899.298421] env[62600]: DEBUG nova.compute.provider_tree [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 899.312191] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222662, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.360715] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 899.360964] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 899.361171] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Deleting the datastore file [datastore2] c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 899.361443] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1363844d-214c-428e-a5ac-dd36036ab28b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.369429] env[62600]: DEBUG oslo_vmware.api [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 899.369429] env[62600]: value = "task-1222663" [ 899.369429] env[62600]: _type = "Task" [ 899.369429] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.378355] env[62600]: DEBUG oslo_vmware.api [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222663, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.389116] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222659, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079451} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.389418] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 899.390244] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2714101d-8960-4bf6-8af8-0a30efb6f25f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.415169] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] d2e43ba2-4ccf-4625-91ba-78a6f2632461/d2e43ba2-4ccf-4625-91ba-78a6f2632461.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 899.416468] env[62600]: DEBUG nova.network.neutron [req-b538ba70-f4f8-42c4-95de-ba046a948dad req-d8f7bb4b-dcf8-46ce-939e-874890bc5097 service nova] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.418308] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da3d16dd-7ee1-402b-bdb5-432b157eef7c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.451679] env[62600]: DEBUG oslo_vmware.api [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Task: {'id': task-1222660, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.37519} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.453197] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 899.453508] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 899.453674] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 899.453929] env[62600]: INFO nova.compute.manager [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Took 1.66 seconds to destroy the instance on the hypervisor. [ 899.454244] env[62600]: DEBUG oslo.service.loopingcall [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 899.454635] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 899.454635] env[62600]: value = "task-1222664" [ 899.454635] env[62600]: _type = "Task" [ 899.454635] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.454952] env[62600]: DEBUG nova.compute.manager [-] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 899.454952] env[62600]: DEBUG nova.network.neutron [-] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 899.467407] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222664, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.525313] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquiring lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.525579] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.525832] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquiring lock "dde6df10-618a-40a8-b33f-efc0ca3a9287-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.526233] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "dde6df10-618a-40a8-b33f-efc0ca3a9287-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.526325] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "dde6df10-618a-40a8-b33f-efc0ca3a9287-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.528804] env[62600]: INFO nova.compute.manager [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Terminating instance [ 899.531091] env[62600]: DEBUG nova.compute.manager [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 899.531281] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 899.532252] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f28a2d91-93c2-4a03-9281-cab7a6cc26c4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.541291] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 899.542211] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6954ba0d-4b25-4367-b684-2daf5ed4f893 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.550249] env[62600]: DEBUG oslo_vmware.api [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for the task: (returnval){ [ 899.550249] env[62600]: value = "task-1222665" [ 899.550249] env[62600]: _type = "Task" [ 899.550249] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.560473] env[62600]: DEBUG oslo_vmware.api [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222665, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.804378] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.585s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.805481] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222662, 'name': Rename_Task, 'duration_secs': 0.451312} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.807693] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 23.088s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.807693] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.807693] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62600) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 899.807693] env[62600]: DEBUG oslo_concurrency.lockutils [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.962s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.807693] env[62600]: DEBUG nova.objects.instance [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Lazy-loading 'resources' on Instance uuid ddc80377-a14e-48d0-ad24-cc6daf7f6850 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.807693] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 899.809262] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccfc728d-aa89-4f0d-879e-1c9d687e8d79 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.812774] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0e35c9da-8a28-496e-9bc3-c2e169b07448 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.820971] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5fc7dea-89be-439d-92e6-0ade6a0ceed3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.826441] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 899.826441] env[62600]: value = "task-1222666" [ 899.826441] env[62600]: _type = "Task" [ 899.826441] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.842171] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2569cd30-6812-46be-8524-4b4d6d3004de {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.850555] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222666, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.858508] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e8b080c-d2e2-4267-ae71-57270cd4b974 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.905438] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180099MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=62600) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 899.905691] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.912747] env[62600]: DEBUG oslo_vmware.api [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222663, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.525683} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.913280] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 899.913615] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 899.913905] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 899.914322] env[62600]: INFO nova.compute.manager [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Took 1.21 seconds to destroy the instance on the hypervisor. [ 899.915534] env[62600]: DEBUG oslo.service.loopingcall [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 899.916069] env[62600]: DEBUG nova.compute.manager [-] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 899.916228] env[62600]: DEBUG nova.network.neutron [-] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 899.940246] env[62600]: DEBUG oslo_concurrency.lockutils [req-b538ba70-f4f8-42c4-95de-ba046a948dad req-d8f7bb4b-dcf8-46ce-939e-874890bc5097 service nova] Releasing lock "refresh_cache-48bbbdaf-496d-4780-b467-6b393257535d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.940853] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquired lock "refresh_cache-48bbbdaf-496d-4780-b467-6b393257535d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.940853] env[62600]: DEBUG nova.network.neutron [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 899.974512] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222664, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.060616] env[62600]: DEBUG oslo_vmware.api [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222665, 'name': PowerOffVM_Task, 'duration_secs': 0.469724} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.061063] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 900.061293] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 900.061624] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-020dcf7c-7c46-434c-9235-12f62b0dc419 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.149527] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 900.149835] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 900.150114] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Deleting the datastore file [datastore1] dde6df10-618a-40a8-b33f-efc0ca3a9287 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 900.150464] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-13bab452-30e2-46d7-ae11-749b0ef779a7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.160837] env[62600]: DEBUG oslo_vmware.api [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for the task: (returnval){ [ 900.160837] env[62600]: value = "task-1222668" [ 900.160837] env[62600]: _type = "Task" [ 900.160837] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.177951] env[62600]: DEBUG oslo_vmware.api [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222668, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.276838] env[62600]: DEBUG nova.compute.manager [req-baddd864-9264-4d85-bf04-912d17b7641b req-e50f6fe7-1069-4085-82d1-4dedad8faea6 service nova] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Received event network-vif-deleted-f39debe1-7447-4681-a72a-83c77b315fd2 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.277067] env[62600]: INFO nova.compute.manager [req-baddd864-9264-4d85-bf04-912d17b7641b req-e50f6fe7-1069-4085-82d1-4dedad8faea6 service nova] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Neutron deleted interface f39debe1-7447-4681-a72a-83c77b315fd2; detaching it from the instance and deleting it from the info cache [ 900.277236] env[62600]: DEBUG nova.network.neutron [req-baddd864-9264-4d85-bf04-912d17b7641b req-e50f6fe7-1069-4085-82d1-4dedad8faea6 service nova] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.308759] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Acquiring lock "c755f398-ddce-47b5-bad3-65a623d90002" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.309014] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Lock "c755f398-ddce-47b5-bad3-65a623d90002" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.343816] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222666, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.464024] env[62600]: DEBUG nova.network.neutron [-] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.473936] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222664, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.497182] env[62600]: DEBUG nova.network.neutron [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 900.657024] env[62600]: DEBUG nova.network.neutron [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Updating instance_info_cache with network_info: [{"id": "f01caa0e-4aff-41e9-a7ed-0511436b131d", "address": "fa:16:3e:2d:75:83", "network": {"id": "353f7a99-4652-4a0d-be26-539c5c42b3d2", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1505004576-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7eca5e197e47759eeea423e280986f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf01caa0e-4a", "ovs_interfaceid": "f01caa0e-4aff-41e9-a7ed-0511436b131d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.657024] env[62600]: DEBUG nova.compute.manager [req-7585b899-9ccf-47a1-9377-e542902a5bdb req-941694c1-a6da-4192-9e19-4f065c801cd9 service nova] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Received event network-vif-deleted-99feed2a-122d-4f5d-890e-0f15c8614b88 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.674858] env[62600]: DEBUG oslo_vmware.api [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Task: {'id': task-1222668, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.302718} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.675164] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 900.675363] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 900.675546] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 900.675721] env[62600]: INFO nova.compute.manager [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Took 1.14 seconds to destroy the instance on the hypervisor. [ 900.675967] env[62600]: DEBUG oslo.service.loopingcall [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.676494] env[62600]: DEBUG nova.compute.manager [-] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 900.676595] env[62600]: DEBUG nova.network.neutron [-] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 900.706017] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad9c98b0-2cc7-4447-bbab-774b021f89db {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.715949] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c0a169-f157-4ed7-a95a-9dcb3ab31e17 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.724952] env[62600]: DEBUG nova.network.neutron [-] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.756767] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fcc32a5-4acb-44ba-8424-c4f9d24b0da8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.766080] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e484a2-d036-4242-b0e8-cd158adbcf71 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.788034] env[62600]: DEBUG nova.compute.provider_tree [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 900.788034] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fdec513d-bf5c-4728-a25c-20ddbc057c1a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.799802] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-414e2327-5979-4422-8c20-52add06b66da {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.811920] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Lock "c755f398-ddce-47b5-bad3-65a623d90002" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.503s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.812526] env[62600]: DEBUG nova.compute.manager [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 900.842389] env[62600]: DEBUG nova.compute.manager [req-baddd864-9264-4d85-bf04-912d17b7641b req-e50f6fe7-1069-4085-82d1-4dedad8faea6 service nova] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Detach interface failed, port_id=f39debe1-7447-4681-a72a-83c77b315fd2, reason: Instance c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 900.848828] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222666, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.970987] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222664, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.974598] env[62600]: INFO nova.compute.manager [-] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Took 1.52 seconds to deallocate network for instance. [ 901.158542] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Releasing lock "refresh_cache-48bbbdaf-496d-4780-b467-6b393257535d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.159067] env[62600]: DEBUG nova.compute.manager [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Instance network_info: |[{"id": "f01caa0e-4aff-41e9-a7ed-0511436b131d", "address": "fa:16:3e:2d:75:83", "network": {"id": "353f7a99-4652-4a0d-be26-539c5c42b3d2", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1505004576-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7eca5e197e47759eeea423e280986f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf01caa0e-4a", "ovs_interfaceid": "f01caa0e-4aff-41e9-a7ed-0511436b131d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 901.159617] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2d:75:83', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd998416-f3d6-4a62-b828-5011063ce76a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f01caa0e-4aff-41e9-a7ed-0511436b131d', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 901.173993] env[62600]: DEBUG oslo.service.loopingcall [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.174880] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 901.175247] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-021aae9c-609b-4405-bea8-224a66ede168 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.199528] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 901.199528] env[62600]: value = "task-1222669" [ 901.199528] env[62600]: _type = "Task" [ 901.199528] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.209223] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222669, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.227871] env[62600]: INFO nova.compute.manager [-] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Took 1.31 seconds to deallocate network for instance. [ 901.318983] env[62600]: DEBUG nova.compute.utils [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 901.319965] env[62600]: DEBUG nova.compute.manager [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 901.320093] env[62600]: DEBUG nova.network.neutron [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 901.343251] env[62600]: DEBUG oslo_vmware.api [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222666, 'name': PowerOnVM_Task, 'duration_secs': 1.058208} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.344314] env[62600]: DEBUG nova.scheduler.client.report [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 100 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 901.344610] env[62600]: DEBUG nova.compute.provider_tree [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 100 to 101 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 901.344805] env[62600]: DEBUG nova.compute.provider_tree [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 901.352973] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 901.353894] env[62600]: INFO nova.compute.manager [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Took 10.14 seconds to spawn the instance on the hypervisor. [ 901.355169] env[62600]: DEBUG nova.compute.manager [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 901.356601] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b44e8b-5591-441c-b39a-32ed5077173e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.376843] env[62600]: DEBUG nova.policy [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f03da8deabe2424e9911287141fe9173', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83082bd5d8e948f380c1542d389b0d53', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 901.473023] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222664, 'name': ReconfigVM_Task, 'duration_secs': 1.666903} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.473023] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Reconfigured VM instance instance-0000004a to attach disk [datastore2] d2e43ba2-4ccf-4625-91ba-78a6f2632461/d2e43ba2-4ccf-4625-91ba-78a6f2632461.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 901.473023] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-58ac983b-42f8-4984-90d9-85310b9241d7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.481304] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.483900] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 901.483900] env[62600]: value = "task-1222670" [ 901.483900] env[62600]: _type = "Task" [ 901.483900] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.494544] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222670, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.696727] env[62600]: DEBUG nova.network.neutron [-] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.713190] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222669, 'name': CreateVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.714321] env[62600]: DEBUG nova.network.neutron [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Successfully created port: abd09e35-cd67-49d8-960d-1713a9907a35 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 901.736317] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.823310] env[62600]: DEBUG nova.compute.manager [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 901.857369] env[62600]: DEBUG oslo_concurrency.lockutils [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.051s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.860667] env[62600]: DEBUG oslo_concurrency.lockutils [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.915s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.862805] env[62600]: INFO nova.compute.claims [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 901.885667] env[62600]: INFO nova.compute.manager [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Took 33.47 seconds to build instance. [ 901.900899] env[62600]: INFO nova.scheduler.client.report [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Deleted allocations for instance ddc80377-a14e-48d0-ad24-cc6daf7f6850 [ 901.900899] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquiring lock "107ed783-10a3-4f07-a809-a5a076c2b904" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.900899] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "107ed783-10a3-4f07-a809-a5a076c2b904" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.900899] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquiring lock "107ed783-10a3-4f07-a809-a5a076c2b904-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.900899] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "107ed783-10a3-4f07-a809-a5a076c2b904-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.900899] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "107ed783-10a3-4f07-a809-a5a076c2b904-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.909845] env[62600]: INFO nova.compute.manager [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Terminating instance [ 901.912087] env[62600]: DEBUG nova.compute.manager [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 901.912087] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 901.913672] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4d3a22-850c-441e-84e3-d31fceb8619c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.926165] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 901.926466] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-643d714a-4012-494d-aea4-480a71507460 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.934493] env[62600]: DEBUG oslo_vmware.api [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 901.934493] env[62600]: value = "task-1222671" [ 901.934493] env[62600]: _type = "Task" [ 901.934493] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.946844] env[62600]: DEBUG oslo_vmware.api [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222671, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.995248] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222670, 'name': Rename_Task, 'duration_secs': 0.206962} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.995950] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 901.997493] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4e1f1d3e-c9fb-4a30-9816-3807b7a6e814 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.008501] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 902.008501] env[62600]: value = "task-1222672" [ 902.008501] env[62600]: _type = "Task" [ 902.008501] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.016992] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222672, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.201083] env[62600]: INFO nova.compute.manager [-] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Took 1.52 seconds to deallocate network for instance. [ 902.216404] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222669, 'name': CreateVM_Task, 'duration_secs': 0.532312} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.216712] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 902.219498] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.219498] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.219498] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 902.220087] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27b0c7de-3ae6-42eb-ab98-a09c4bf2f7d9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.227565] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 902.227565] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e421f9-7d0d-3d0b-6ec6-866040418590" [ 902.227565] env[62600]: _type = "Task" [ 902.227565] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.239729] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e421f9-7d0d-3d0b-6ec6-866040418590, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.390209] env[62600]: DEBUG oslo_concurrency.lockutils [None req-28593388-13e1-4c64-a0b3-6f33255f2f67 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "d627a701-77cd-4a1d-9e52-0fc4adc93391" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.993s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.415766] env[62600]: DEBUG nova.compute.manager [req-29f9593f-c0c2-4220-b2c6-5a5a050a7986 req-38f375d2-be18-4c73-b8e6-4a746cb43ccf service nova] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Received event network-vif-deleted-6832182f-fe16-4367-a4bc-53d7bd8f6033 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.416692] env[62600]: DEBUG oslo_concurrency.lockutils [None req-71f5f067-c18d-4b13-b230-d13662910fb1 tempest-ServerMetadataTestJSON-1129518119 tempest-ServerMetadataTestJSON-1129518119-project-member] Lock "ddc80377-a14e-48d0-ad24-cc6daf7f6850" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.999s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.446544] env[62600]: DEBUG oslo_vmware.api [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222671, 'name': PowerOffVM_Task, 'duration_secs': 0.28776} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.446653] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 902.446853] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 902.447139] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e6e1ecab-8d67-499d-9da4-fa8b7cef5616 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.518632] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222672, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.537016] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 902.537016] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 902.537016] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Deleting the datastore file [datastore1] 107ed783-10a3-4f07-a809-a5a076c2b904 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 902.537016] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-615252ee-c1f7-4aa3-9c33-23e737b884f3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.543565] env[62600]: DEBUG oslo_vmware.api [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 902.543565] env[62600]: value = "task-1222674" [ 902.543565] env[62600]: _type = "Task" [ 902.543565] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.552570] env[62600]: DEBUG oslo_vmware.api [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222674, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.716940] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.740355] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e421f9-7d0d-3d0b-6ec6-866040418590, 'name': SearchDatastore_Task, 'duration_secs': 0.014113} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.740680] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.740930] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 902.741469] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.741688] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.742086] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 902.742395] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-50f60a45-abef-4ac8-86f7-e01dc15c2820 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.753647] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 902.753943] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 902.755116] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ac1c92a-d72b-4e44-9ab5-7fab3c8d1d77 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.763720] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 902.763720] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5251e78f-4939-8530-85dd-a3c7c444a0d9" [ 902.763720] env[62600]: _type = "Task" [ 902.763720] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.773397] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5251e78f-4939-8530-85dd-a3c7c444a0d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.837138] env[62600]: DEBUG nova.compute.manager [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 902.863530] env[62600]: DEBUG nova.virt.hardware [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 902.864254] env[62600]: DEBUG nova.virt.hardware [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 902.864607] env[62600]: DEBUG nova.virt.hardware [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 902.864977] env[62600]: DEBUG nova.virt.hardware [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 902.865266] env[62600]: DEBUG nova.virt.hardware [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 902.865547] env[62600]: DEBUG nova.virt.hardware [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 902.865885] env[62600]: DEBUG nova.virt.hardware [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 902.866264] env[62600]: DEBUG nova.virt.hardware [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 902.866558] env[62600]: DEBUG nova.virt.hardware [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 902.866927] env[62600]: DEBUG nova.virt.hardware [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 902.867515] env[62600]: DEBUG nova.virt.hardware [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 902.868944] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d27d5d6-ee6d-44ce-bf81-107243e341f4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.884870] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2277d117-83ec-4de6-b0f7-db58723f99f8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.019917] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222672, 'name': PowerOnVM_Task, 'duration_secs': 0.682455} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.020314] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 903.020528] env[62600]: INFO nova.compute.manager [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Took 9.34 seconds to spawn the instance on the hypervisor. [ 903.020714] env[62600]: DEBUG nova.compute.manager [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 903.021511] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-481a0cbf-c0dd-47a9-8f8a-745443e3d014 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.057995] env[62600]: DEBUG oslo_vmware.api [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222674, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.220302} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.060827] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 903.061200] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 903.061506] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 903.061845] env[62600]: INFO nova.compute.manager [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Took 1.15 seconds to destroy the instance on the hypervisor. [ 903.062226] env[62600]: DEBUG oslo.service.loopingcall [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.062700] env[62600]: DEBUG nova.compute.manager [-] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 903.062900] env[62600]: DEBUG nova.network.neutron [-] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 903.171108] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aec5494e-7c7f-4b81-8eb1-dcf798d1f1c1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.180107] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d130d27-874c-4e1c-9293-cd5f1346c443 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.218442] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293b6b7e-b216-4218-ac08-1174a3dc0009 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.222316] env[62600]: INFO nova.compute.manager [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Rebuilding instance [ 903.233350] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dbcdf49-3f2a-44d0-a002-dc358f69112d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.251112] env[62600]: DEBUG nova.compute.provider_tree [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.278430] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5251e78f-4939-8530-85dd-a3c7c444a0d9, 'name': SearchDatastore_Task, 'duration_secs': 0.02513} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.284627] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d71ebdf-9635-490c-bb5f-c1add21e6527 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.286920] env[62600]: DEBUG nova.compute.manager [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 903.288534] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03e24b9b-8f76-40ff-a9fd-e7180181eb91 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.302998] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 903.302998] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]524f3157-b910-c37d-b060-1b2e4f378628" [ 903.302998] env[62600]: _type = "Task" [ 903.302998] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.316376] env[62600]: DEBUG nova.network.neutron [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Successfully updated port: abd09e35-cd67-49d8-960d-1713a9907a35 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 903.317278] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]524f3157-b910-c37d-b060-1b2e4f378628, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.554622] env[62600]: INFO nova.compute.manager [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Took 31.07 seconds to build instance. [ 903.759965] env[62600]: DEBUG nova.scheduler.client.report [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.805596] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 903.806534] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-889ba7d1-b56a-40f4-9056-3963542de771 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.821114] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Acquiring lock "refresh_cache-4eaa22b9-5834-4425-82a9-d16e316f3a52" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.821313] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Acquired lock "refresh_cache-4eaa22b9-5834-4425-82a9-d16e316f3a52" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.821470] env[62600]: DEBUG nova.network.neutron [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 903.823241] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]524f3157-b910-c37d-b060-1b2e4f378628, 'name': SearchDatastore_Task, 'duration_secs': 0.029448} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.824926] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.825272] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 48bbbdaf-496d-4780-b467-6b393257535d/48bbbdaf-496d-4780-b467-6b393257535d.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 903.825654] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 903.825654] env[62600]: value = "task-1222675" [ 903.825654] env[62600]: _type = "Task" [ 903.825654] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.825901] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c9d98e0e-8576-4644-b4ea-09c108eec905 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.846838] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222675, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.848545] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 903.848545] env[62600]: value = "task-1222676" [ 903.848545] env[62600]: _type = "Task" [ 903.848545] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.863050] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222676, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.889097] env[62600]: DEBUG nova.compute.manager [req-b8e3121c-8081-4847-91c8-5356beb0022b req-79965a2b-8cef-475e-a6cd-1a8aefa3bd25 service nova] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Received event network-vif-deleted-420267c8-eb2f-4f1f-b041-8eca21b9118b {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 903.889097] env[62600]: INFO nova.compute.manager [req-b8e3121c-8081-4847-91c8-5356beb0022b req-79965a2b-8cef-475e-a6cd-1a8aefa3bd25 service nova] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Neutron deleted interface 420267c8-eb2f-4f1f-b041-8eca21b9118b; detaching it from the instance and deleting it from the info cache [ 903.889097] env[62600]: DEBUG nova.network.neutron [req-b8e3121c-8081-4847-91c8-5356beb0022b req-79965a2b-8cef-475e-a6cd-1a8aefa3bd25 service nova] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.056482] env[62600]: DEBUG oslo_concurrency.lockutils [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Acquiring lock "c2f38d56-72fa-4bf0-a2cb-13d666571466" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.056717] env[62600]: DEBUG oslo_concurrency.lockutils [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Lock "c2f38d56-72fa-4bf0-a2cb-13d666571466" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.058506] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "d2e43ba2-4ccf-4625-91ba-78a6f2632461" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.588s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.180090] env[62600]: DEBUG nova.network.neutron [-] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.264936] env[62600]: DEBUG oslo_concurrency.lockutils [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.265599] env[62600]: DEBUG nova.compute.manager [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 904.268650] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.517s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.268960] env[62600]: DEBUG nova.objects.instance [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lazy-loading 'resources' on Instance uuid bd9b7351-01a2-429d-a860-d9ff6855eefc {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.347755] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222675, 'name': PowerOffVM_Task, 'duration_secs': 0.206775} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.349466] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 904.349466] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 904.350229] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ac9151-2c50-4d20-99ec-b647006ca58b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.364922] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222676, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.367792] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 904.368280] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-913b57c4-6d86-46e8-b86a-60cd70f3940f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.394027] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-310ecb4b-b569-499b-8cb4-e065290ee4cf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.399628] env[62600]: DEBUG nova.network.neutron [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 904.409720] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db357d50-d3c4-4dab-b800-ea7aeb2cffbc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.446052] env[62600]: DEBUG nova.compute.manager [req-b8e3121c-8081-4847-91c8-5356beb0022b req-79965a2b-8cef-475e-a6cd-1a8aefa3bd25 service nova] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Detach interface failed, port_id=420267c8-eb2f-4f1f-b041-8eca21b9118b, reason: Instance 107ed783-10a3-4f07-a809-a5a076c2b904 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 904.449681] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 904.450055] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 904.454025] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleting the datastore file [datastore2] d627a701-77cd-4a1d-9e52-0fc4adc93391 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 904.454025] env[62600]: DEBUG nova.compute.manager [req-9093376e-72c6-4a98-9087-8cc99f169687 req-f47bb2d5-1144-4dd9-92e9-4f065365e7f2 service nova] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Received event network-vif-plugged-abd09e35-cd67-49d8-960d-1713a9907a35 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 904.454025] env[62600]: DEBUG oslo_concurrency.lockutils [req-9093376e-72c6-4a98-9087-8cc99f169687 req-f47bb2d5-1144-4dd9-92e9-4f065365e7f2 service nova] Acquiring lock "4eaa22b9-5834-4425-82a9-d16e316f3a52-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.454025] env[62600]: DEBUG oslo_concurrency.lockutils [req-9093376e-72c6-4a98-9087-8cc99f169687 req-f47bb2d5-1144-4dd9-92e9-4f065365e7f2 service nova] Lock "4eaa22b9-5834-4425-82a9-d16e316f3a52-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.454025] env[62600]: DEBUG oslo_concurrency.lockutils [req-9093376e-72c6-4a98-9087-8cc99f169687 req-f47bb2d5-1144-4dd9-92e9-4f065365e7f2 service nova] Lock "4eaa22b9-5834-4425-82a9-d16e316f3a52-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.454025] env[62600]: DEBUG nova.compute.manager [req-9093376e-72c6-4a98-9087-8cc99f169687 req-f47bb2d5-1144-4dd9-92e9-4f065365e7f2 service nova] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] No waiting events found dispatching network-vif-plugged-abd09e35-cd67-49d8-960d-1713a9907a35 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 904.454025] env[62600]: WARNING nova.compute.manager [req-9093376e-72c6-4a98-9087-8cc99f169687 req-f47bb2d5-1144-4dd9-92e9-4f065365e7f2 service nova] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Received unexpected event network-vif-plugged-abd09e35-cd67-49d8-960d-1713a9907a35 for instance with vm_state building and task_state spawning. [ 904.454025] env[62600]: DEBUG nova.compute.manager [req-9093376e-72c6-4a98-9087-8cc99f169687 req-f47bb2d5-1144-4dd9-92e9-4f065365e7f2 service nova] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Received event network-changed-abd09e35-cd67-49d8-960d-1713a9907a35 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 904.454025] env[62600]: DEBUG nova.compute.manager [req-9093376e-72c6-4a98-9087-8cc99f169687 req-f47bb2d5-1144-4dd9-92e9-4f065365e7f2 service nova] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Refreshing instance network info cache due to event network-changed-abd09e35-cd67-49d8-960d-1713a9907a35. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 904.454025] env[62600]: DEBUG oslo_concurrency.lockutils [req-9093376e-72c6-4a98-9087-8cc99f169687 req-f47bb2d5-1144-4dd9-92e9-4f065365e7f2 service nova] Acquiring lock "refresh_cache-4eaa22b9-5834-4425-82a9-d16e316f3a52" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.454713] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-150739ef-a83f-43cc-a021-755ffb108128 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.463952] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 904.463952] env[62600]: value = "task-1222678" [ 904.463952] env[62600]: _type = "Task" [ 904.463952] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.475273] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222678, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.559133] env[62600]: DEBUG nova.compute.manager [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 904.683102] env[62600]: INFO nova.compute.manager [-] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Took 1.62 seconds to deallocate network for instance. [ 904.685896] env[62600]: DEBUG nova.network.neutron [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Updating instance_info_cache with network_info: [{"id": "abd09e35-cd67-49d8-960d-1713a9907a35", "address": "fa:16:3e:1e:cc:73", "network": {"id": "9e605fbb-f192-40f4-b9c0-185103dd70df", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1826190130-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83082bd5d8e948f380c1542d389b0d53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapabd09e35-cd", "ovs_interfaceid": "abd09e35-cd67-49d8-960d-1713a9907a35", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.779143] env[62600]: DEBUG nova.compute.utils [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 904.779143] env[62600]: DEBUG nova.compute.manager [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 904.779143] env[62600]: DEBUG nova.network.neutron [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 904.860062] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222676, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.719429} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.862808] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 48bbbdaf-496d-4780-b467-6b393257535d/48bbbdaf-496d-4780-b467-6b393257535d.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 904.863068] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 904.863568] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-01924d32-8314-4c0c-afa2-78235332c82e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.871993] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 904.871993] env[62600]: value = "task-1222679" [ 904.871993] env[62600]: _type = "Task" [ 904.871993] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.893167] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222679, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.920012] env[62600]: DEBUG nova.policy [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34319d68370c48ef93fa24a9d9132407', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bab74e3ae78248909dbcd483abbb8da2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 904.979160] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222678, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.079420] env[62600]: DEBUG oslo_concurrency.lockutils [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.126978] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1caec3c1-fbc7-46d6-96b5-2278bb5bacb2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.135132] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d54c91-6332-4624-843f-08acf127983e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.170353] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e1cc344-cdde-4681-aafc-ad6caecb040b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.178765] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f00de99-a93a-4b75-a70e-5295e083bec2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.193743] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Releasing lock "refresh_cache-4eaa22b9-5834-4425-82a9-d16e316f3a52" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.194144] env[62600]: DEBUG nova.compute.manager [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Instance network_info: |[{"id": "abd09e35-cd67-49d8-960d-1713a9907a35", "address": "fa:16:3e:1e:cc:73", "network": {"id": "9e605fbb-f192-40f4-b9c0-185103dd70df", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1826190130-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83082bd5d8e948f380c1542d389b0d53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapabd09e35-cd", "ovs_interfaceid": "abd09e35-cd67-49d8-960d-1713a9907a35", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 905.195027] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.195575] env[62600]: DEBUG nova.compute.provider_tree [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.196726] env[62600]: DEBUG oslo_concurrency.lockutils [req-9093376e-72c6-4a98-9087-8cc99f169687 req-f47bb2d5-1144-4dd9-92e9-4f065365e7f2 service nova] Acquired lock "refresh_cache-4eaa22b9-5834-4425-82a9-d16e316f3a52" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.196919] env[62600]: DEBUG nova.network.neutron [req-9093376e-72c6-4a98-9087-8cc99f169687 req-f47bb2d5-1144-4dd9-92e9-4f065365e7f2 service nova] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Refreshing network info cache for port abd09e35-cd67-49d8-960d-1713a9907a35 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 905.198018] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:cc:73', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '459b8c74-0aa6-42b6-996a-42b1c5d7e5c6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'abd09e35-cd67-49d8-960d-1713a9907a35', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 905.205346] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Creating folder: Project (83082bd5d8e948f380c1542d389b0d53). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 905.206023] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ce30763a-dbbe-4e50-b15f-7655ade5f889 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.219297] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Created folder: Project (83082bd5d8e948f380c1542d389b0d53) in parent group-v264198. [ 905.219506] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Creating folder: Instances. Parent ref: group-v264320. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 905.219777] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bd1c0169-c1c9-4427-ae1f-7f1062c0dbc6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.232090] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Created folder: Instances in parent group-v264320. [ 905.232090] env[62600]: DEBUG oslo.service.loopingcall [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.232090] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 905.232090] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-be40f39a-8610-4d18-8b5c-d998a5f8263e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.252970] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 905.252970] env[62600]: value = "task-1222682" [ 905.252970] env[62600]: _type = "Task" [ 905.252970] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.262266] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222682, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.283170] env[62600]: DEBUG nova.compute.manager [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 905.298131] env[62600]: DEBUG nova.network.neutron [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Successfully created port: a0340049-6048-4c62-8468-1af1b2551a37 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 905.385418] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222679, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074641} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.385734] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 905.386839] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1601b6f7-aaf0-42e4-8478-88550a2e281e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.426675] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] 48bbbdaf-496d-4780-b467-6b393257535d/48bbbdaf-496d-4780-b467-6b393257535d.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 905.427278] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36f601ba-6883-43e9-a78f-14955f9bf4f4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.449474] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 905.449474] env[62600]: value = "task-1222683" [ 905.449474] env[62600]: _type = "Task" [ 905.449474] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.463579] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222683, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.477316] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222678, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.735081} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.477660] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 905.477911] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 905.478197] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 905.710071] env[62600]: DEBUG nova.scheduler.client.report [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.764082] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222682, 'name': CreateVM_Task, 'duration_secs': 0.381302} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.764365] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 905.765716] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.765716] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.765869] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 905.766099] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8784cc2e-84cb-426d-82e4-b70e634d6443 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.772721] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Waiting for the task: (returnval){ [ 905.772721] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52845cb4-999f-c5a4-428d-5d0ba50cd64f" [ 905.772721] env[62600]: _type = "Task" [ 905.772721] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.784843] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52845cb4-999f-c5a4-428d-5d0ba50cd64f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.961368] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222683, 'name': ReconfigVM_Task, 'duration_secs': 0.317777} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.961731] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Reconfigured VM instance instance-0000004b to attach disk [datastore2] 48bbbdaf-496d-4780-b467-6b393257535d/48bbbdaf-496d-4780-b467-6b393257535d.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.962408] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-785076b7-aa16-4b98-8fec-c2b8b8db20c9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.974256] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 905.974256] env[62600]: value = "task-1222684" [ 905.974256] env[62600]: _type = "Task" [ 905.974256] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.984513] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222684, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.045272] env[62600]: DEBUG nova.network.neutron [req-9093376e-72c6-4a98-9087-8cc99f169687 req-f47bb2d5-1144-4dd9-92e9-4f065365e7f2 service nova] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Updated VIF entry in instance network info cache for port abd09e35-cd67-49d8-960d-1713a9907a35. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 906.045272] env[62600]: DEBUG nova.network.neutron [req-9093376e-72c6-4a98-9087-8cc99f169687 req-f47bb2d5-1144-4dd9-92e9-4f065365e7f2 service nova] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Updating instance_info_cache with network_info: [{"id": "abd09e35-cd67-49d8-960d-1713a9907a35", "address": "fa:16:3e:1e:cc:73", "network": {"id": "9e605fbb-f192-40f4-b9c0-185103dd70df", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1826190130-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "83082bd5d8e948f380c1542d389b0d53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "459b8c74-0aa6-42b6-996a-42b1c5d7e5c6", "external-id": "nsx-vlan-transportzone-467", "segmentation_id": 467, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapabd09e35-cd", "ovs_interfaceid": "abd09e35-cd67-49d8-960d-1713a9907a35", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.215735] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.947s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.218311] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.924s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.219942] env[62600]: INFO nova.compute.claims [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 906.242420] env[62600]: INFO nova.scheduler.client.report [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Deleted allocations for instance bd9b7351-01a2-429d-a860-d9ff6855eefc [ 906.287570] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52845cb4-999f-c5a4-428d-5d0ba50cd64f, 'name': SearchDatastore_Task, 'duration_secs': 0.013448} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.288118] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.288482] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 906.288845] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.289177] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.289487] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 906.289925] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-657f4044-9e48-485e-9469-7377eb8ee9c5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.292846] env[62600]: DEBUG nova.compute.manager [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 906.304681] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 906.306794] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 906.306794] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18fe8877-1af6-42e2-9324-f2c74697cf27 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.317444] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Waiting for the task: (returnval){ [ 906.317444] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525a363b-3f18-829e-b684-2108f51191ad" [ 906.317444] env[62600]: _type = "Task" [ 906.317444] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.319786] env[62600]: DEBUG nova.virt.hardware [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 906.320031] env[62600]: DEBUG nova.virt.hardware [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 906.320205] env[62600]: DEBUG nova.virt.hardware [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 906.320403] env[62600]: DEBUG nova.virt.hardware [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 906.320559] env[62600]: DEBUG nova.virt.hardware [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 906.320716] env[62600]: DEBUG nova.virt.hardware [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 906.320937] env[62600]: DEBUG nova.virt.hardware [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 906.321121] env[62600]: DEBUG nova.virt.hardware [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 906.321297] env[62600]: DEBUG nova.virt.hardware [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 906.321469] env[62600]: DEBUG nova.virt.hardware [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 906.321651] env[62600]: DEBUG nova.virt.hardware [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 906.323283] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a602e6-c23c-470b-bf3a-883a90d58909 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.337930] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8933abeb-f8c7-4e16-bfbc-aafbd48be049 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.343314] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525a363b-3f18-829e-b684-2108f51191ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.486255] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222684, 'name': Rename_Task, 'duration_secs': 0.265529} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.486557] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 906.486811] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f5b6baa-302a-47ef-8e9b-ea5a490563f6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.494417] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 906.494417] env[62600]: value = "task-1222685" [ 906.494417] env[62600]: _type = "Task" [ 906.494417] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.504051] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222685, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.513817] env[62600]: DEBUG nova.virt.hardware [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 906.514097] env[62600]: DEBUG nova.virt.hardware [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 906.514270] env[62600]: DEBUG nova.virt.hardware [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 906.514464] env[62600]: DEBUG nova.virt.hardware [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 906.514621] env[62600]: DEBUG nova.virt.hardware [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 906.514789] env[62600]: DEBUG nova.virt.hardware [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 906.515028] env[62600]: DEBUG nova.virt.hardware [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 906.515205] env[62600]: DEBUG nova.virt.hardware [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 906.515384] env[62600]: DEBUG nova.virt.hardware [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 906.515554] env[62600]: DEBUG nova.virt.hardware [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 906.515736] env[62600]: DEBUG nova.virt.hardware [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 906.516620] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b59b40-bb1d-4365-b255-7c5a04ddbcd3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.525432] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21844219-7005-402d-84bb-0de72daf3f50 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.541451] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:16:7c:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43798f54-0c07-4417-a23f-58bb6b7e204b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '92bc4652-839c-47b3-bcf5-2cc0c43d5537', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 906.549521] env[62600]: DEBUG oslo.service.loopingcall [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 906.550376] env[62600]: DEBUG oslo_concurrency.lockutils [req-9093376e-72c6-4a98-9087-8cc99f169687 req-f47bb2d5-1144-4dd9-92e9-4f065365e7f2 service nova] Releasing lock "refresh_cache-4eaa22b9-5834-4425-82a9-d16e316f3a52" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.550791] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 906.551093] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b383b5c3-a328-42a8-9175-15d6eb46eb0f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.572027] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 906.572027] env[62600]: value = "task-1222686" [ 906.572027] env[62600]: _type = "Task" [ 906.572027] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.580702] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222686, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.750870] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e0701f71-06a1-42fd-b596-aeba9e1303b5 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "bd9b7351-01a2-429d-a860-d9ff6855eefc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.418s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.835419] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525a363b-3f18-829e-b684-2108f51191ad, 'name': SearchDatastore_Task, 'duration_secs': 0.033418} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.835631] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53b6be44-8e50-4dc6-98de-36051b0cc952 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.843076] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Waiting for the task: (returnval){ [ 906.843076] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52474c8d-bfa5-6285-d607-1efe7e49536d" [ 906.843076] env[62600]: _type = "Task" [ 906.843076] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.853354] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52474c8d-bfa5-6285-d607-1efe7e49536d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.972263] env[62600]: DEBUG nova.compute.manager [req-b6016085-591c-47ee-877b-6d04fc514abc req-913b68ee-987d-46b3-b361-1c22310dd840 service nova] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Received event network-vif-plugged-a0340049-6048-4c62-8468-1af1b2551a37 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.973062] env[62600]: DEBUG oslo_concurrency.lockutils [req-b6016085-591c-47ee-877b-6d04fc514abc req-913b68ee-987d-46b3-b361-1c22310dd840 service nova] Acquiring lock "043ac602-6c4a-4275-b019-5d717579e7c0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.973337] env[62600]: DEBUG oslo_concurrency.lockutils [req-b6016085-591c-47ee-877b-6d04fc514abc req-913b68ee-987d-46b3-b361-1c22310dd840 service nova] Lock "043ac602-6c4a-4275-b019-5d717579e7c0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.973772] env[62600]: DEBUG oslo_concurrency.lockutils [req-b6016085-591c-47ee-877b-6d04fc514abc req-913b68ee-987d-46b3-b361-1c22310dd840 service nova] Lock "043ac602-6c4a-4275-b019-5d717579e7c0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.974149] env[62600]: DEBUG nova.compute.manager [req-b6016085-591c-47ee-877b-6d04fc514abc req-913b68ee-987d-46b3-b361-1c22310dd840 service nova] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] No waiting events found dispatching network-vif-plugged-a0340049-6048-4c62-8468-1af1b2551a37 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 906.974531] env[62600]: WARNING nova.compute.manager [req-b6016085-591c-47ee-877b-6d04fc514abc req-913b68ee-987d-46b3-b361-1c22310dd840 service nova] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Received unexpected event network-vif-plugged-a0340049-6048-4c62-8468-1af1b2551a37 for instance with vm_state building and task_state spawning. [ 907.005677] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222685, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.082742] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222686, 'name': CreateVM_Task, 'duration_secs': 0.459186} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.082913] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 907.083615] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.083787] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.084147] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 907.084401] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf45f60c-d465-4536-a57f-b48718b9ab71 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.090569] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 907.090569] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5210982a-9ffb-7d51-2acb-122bec0a60b4" [ 907.090569] env[62600]: _type = "Task" [ 907.090569] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.094631] env[62600]: DEBUG nova.network.neutron [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Successfully updated port: a0340049-6048-4c62-8468-1af1b2551a37 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 907.108216] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5210982a-9ffb-7d51-2acb-122bec0a60b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.355112] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52474c8d-bfa5-6285-d607-1efe7e49536d, 'name': SearchDatastore_Task, 'duration_secs': 0.016627} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.357632] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.357904] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 4eaa22b9-5834-4425-82a9-d16e316f3a52/4eaa22b9-5834-4425-82a9-d16e316f3a52.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 907.359177] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dd10ed44-87a7-49a7-81ab-35a6e6c402d5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.367297] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Waiting for the task: (returnval){ [ 907.367297] env[62600]: value = "task-1222687" [ 907.367297] env[62600]: _type = "Task" [ 907.367297] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.376738] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': task-1222687, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.505595] env[62600]: DEBUG oslo_vmware.api [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222685, 'name': PowerOnVM_Task, 'duration_secs': 0.687317} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.508293] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 907.508512] env[62600]: INFO nova.compute.manager [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Took 10.69 seconds to spawn the instance on the hypervisor. [ 907.508694] env[62600]: DEBUG nova.compute.manager [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 907.510167] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f66d43f-29cd-40d7-9eda-bee6c5cdd236 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.540066] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6418ded-3ef3-446f-a9d7-5b68423e7fb1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.550147] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8430a52-4ccb-47de-b79d-a1b544719579 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.584062] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4cf7d07-951b-4d50-984c-891c301a33d1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.597914] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4498f200-68d1-4c0d-93ee-4ef76d7031e8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.609875] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5210982a-9ffb-7d51-2acb-122bec0a60b4, 'name': SearchDatastore_Task, 'duration_secs': 0.018873} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.615777] env[62600]: DEBUG oslo_concurrency.lockutils [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "refresh_cache-043ac602-6c4a-4275-b019-5d717579e7c0" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.615925] env[62600]: DEBUG oslo_concurrency.lockutils [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired lock "refresh_cache-043ac602-6c4a-4275-b019-5d717579e7c0" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.616089] env[62600]: DEBUG nova.network.neutron [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 907.620085] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.620085] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 907.620085] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.620270] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.620342] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 907.621265] env[62600]: DEBUG nova.compute.provider_tree [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.622422] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9b6352b9-ffdb-4618-917e-daaa09217f4e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.633137] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 907.633334] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 907.634570] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2fe7d58a-9640-4672-a60d-c89e2bb056d6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.641372] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 907.641372] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b2adb5-c2b0-bf66-a084-6d0a5d3c86f1" [ 907.641372] env[62600]: _type = "Task" [ 907.641372] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.650100] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b2adb5-c2b0-bf66-a084-6d0a5d3c86f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.883549] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': task-1222687, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.031126] env[62600]: INFO nova.compute.manager [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Took 35.51 seconds to build instance. [ 908.126843] env[62600]: DEBUG nova.scheduler.client.report [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 908.153639] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b2adb5-c2b0-bf66-a084-6d0a5d3c86f1, 'name': SearchDatastore_Task, 'duration_secs': 0.010227} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.154553] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20d3a588-4fd7-44e4-9051-8e65a931c321 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.161324] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 908.161324] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52406354-b07c-8977-3037-ebbeaf3b1f35" [ 908.161324] env[62600]: _type = "Task" [ 908.161324] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.170016] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52406354-b07c-8977-3037-ebbeaf3b1f35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.175757] env[62600]: DEBUG nova.network.neutron [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 908.379332] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': task-1222687, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.739218} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.379646] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 4eaa22b9-5834-4425-82a9-d16e316f3a52/4eaa22b9-5834-4425-82a9-d16e316f3a52.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 908.379901] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 908.380149] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1944c892-f39b-4308-8690-fa0b255936b5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.389158] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Waiting for the task: (returnval){ [ 908.389158] env[62600]: value = "task-1222688" [ 908.389158] env[62600]: _type = "Task" [ 908.389158] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.398881] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': task-1222688, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.435668] env[62600]: DEBUG nova.network.neutron [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Updating instance_info_cache with network_info: [{"id": "a0340049-6048-4c62-8468-1af1b2551a37", "address": "fa:16:3e:32:5b:6a", "network": {"id": "5ef67e8a-5613-4cdb-9f35-0a18720657ff", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1644754500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bab74e3ae78248909dbcd483abbb8da2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0340049-60", "ovs_interfaceid": "a0340049-6048-4c62-8468-1af1b2551a37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.534160] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22dd0a24-52d3-48dd-864c-75bfdd28b41e tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "48bbbdaf-496d-4780-b467-6b393257535d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.030s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.632920] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.414s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.633028] env[62600]: DEBUG nova.compute.manager [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 908.636205] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.718s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.640015] env[62600]: INFO nova.compute.claims [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 908.683645] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52406354-b07c-8977-3037-ebbeaf3b1f35, 'name': SearchDatastore_Task, 'duration_secs': 0.03652} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.684362] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.684799] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] d627a701-77cd-4a1d-9e52-0fc4adc93391/d627a701-77cd-4a1d-9e52-0fc4adc93391.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 908.685434] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3f16a688-73f4-4e16-be5b-2a280ba2ee02 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.697172] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 908.697172] env[62600]: value = "task-1222689" [ 908.697172] env[62600]: _type = "Task" [ 908.697172] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.709314] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222689, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.726921] env[62600]: DEBUG oslo_concurrency.lockutils [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "d2e43ba2-4ccf-4625-91ba-78a6f2632461" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.727253] env[62600]: DEBUG oslo_concurrency.lockutils [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "d2e43ba2-4ccf-4625-91ba-78a6f2632461" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.727489] env[62600]: DEBUG oslo_concurrency.lockutils [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "d2e43ba2-4ccf-4625-91ba-78a6f2632461-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.727689] env[62600]: DEBUG oslo_concurrency.lockutils [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "d2e43ba2-4ccf-4625-91ba-78a6f2632461-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.727902] env[62600]: DEBUG oslo_concurrency.lockutils [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "d2e43ba2-4ccf-4625-91ba-78a6f2632461-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.731843] env[62600]: INFO nova.compute.manager [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Terminating instance [ 908.733810] env[62600]: DEBUG nova.compute.manager [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 908.734033] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 908.734894] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ef8df2-f459-483a-bf33-b38415938a93 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.743425] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 908.743668] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9482d38e-5546-46cf-aa6e-ee20bc064031 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.751962] env[62600]: DEBUG oslo_vmware.api [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 908.751962] env[62600]: value = "task-1222690" [ 908.751962] env[62600]: _type = "Task" [ 908.751962] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.760628] env[62600]: DEBUG oslo_vmware.api [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222690, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.899304] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': task-1222688, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072694} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.899571] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 908.900392] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49937d3e-04b5-478c-a7bf-ae0f8fa5d2a3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.923797] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] 4eaa22b9-5834-4425-82a9-d16e316f3a52/4eaa22b9-5834-4425-82a9-d16e316f3a52.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 908.924236] env[62600]: DEBUG oslo_concurrency.lockutils [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "48bbbdaf-496d-4780-b467-6b393257535d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.924468] env[62600]: DEBUG oslo_concurrency.lockutils [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "48bbbdaf-496d-4780-b467-6b393257535d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.924671] env[62600]: DEBUG oslo_concurrency.lockutils [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "48bbbdaf-496d-4780-b467-6b393257535d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.924863] env[62600]: DEBUG oslo_concurrency.lockutils [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "48bbbdaf-496d-4780-b467-6b393257535d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.925052] env[62600]: DEBUG oslo_concurrency.lockutils [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "48bbbdaf-496d-4780-b467-6b393257535d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.926806] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a6880e4-cfdd-4e06-a88c-452137f5aeb0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.942878] env[62600]: INFO nova.compute.manager [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Terminating instance [ 908.945195] env[62600]: DEBUG oslo_concurrency.lockutils [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Releasing lock "refresh_cache-043ac602-6c4a-4275-b019-5d717579e7c0" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.945520] env[62600]: DEBUG nova.compute.manager [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Instance network_info: |[{"id": "a0340049-6048-4c62-8468-1af1b2551a37", "address": "fa:16:3e:32:5b:6a", "network": {"id": "5ef67e8a-5613-4cdb-9f35-0a18720657ff", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1644754500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bab74e3ae78248909dbcd483abbb8da2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0340049-60", "ovs_interfaceid": "a0340049-6048-4c62-8468-1af1b2551a37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 908.946335] env[62600]: DEBUG nova.compute.manager [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 908.946588] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 908.947105] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:32:5b:6a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3734b156-0f7d-4721-b23c-d000412ec2eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a0340049-6048-4c62-8468-1af1b2551a37', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 908.955646] env[62600]: DEBUG oslo.service.loopingcall [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 908.956985] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-474e4eb0-65f3-4979-b5c4-f64bc316b0ca {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.960250] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 908.960880] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bfdeae81-9b7d-4864-a45c-471a02ecea0c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.978396] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Waiting for the task: (returnval){ [ 908.978396] env[62600]: value = "task-1222691" [ 908.978396] env[62600]: _type = "Task" [ 908.978396] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.984173] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 908.984800] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-63e8676f-806b-40b6-b4b3-a32181ac4c50 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.987802] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 908.987802] env[62600]: value = "task-1222692" [ 908.987802] env[62600]: _type = "Task" [ 908.987802] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.991184] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': task-1222691, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.995451] env[62600]: DEBUG oslo_vmware.api [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 908.995451] env[62600]: value = "task-1222693" [ 908.995451] env[62600]: _type = "Task" [ 908.995451] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.002146] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222692, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.008432] env[62600]: DEBUG nova.compute.manager [req-5bb2717e-15fe-4d56-9d33-05d6e8d22b1c req-6f681a1c-f14f-48a5-8cae-c088d7f5e61f service nova] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Received event network-changed-a0340049-6048-4c62-8468-1af1b2551a37 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.008432] env[62600]: DEBUG nova.compute.manager [req-5bb2717e-15fe-4d56-9d33-05d6e8d22b1c req-6f681a1c-f14f-48a5-8cae-c088d7f5e61f service nova] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Refreshing instance network info cache due to event network-changed-a0340049-6048-4c62-8468-1af1b2551a37. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 909.008432] env[62600]: DEBUG oslo_concurrency.lockutils [req-5bb2717e-15fe-4d56-9d33-05d6e8d22b1c req-6f681a1c-f14f-48a5-8cae-c088d7f5e61f service nova] Acquiring lock "refresh_cache-043ac602-6c4a-4275-b019-5d717579e7c0" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.008778] env[62600]: DEBUG oslo_concurrency.lockutils [req-5bb2717e-15fe-4d56-9d33-05d6e8d22b1c req-6f681a1c-f14f-48a5-8cae-c088d7f5e61f service nova] Acquired lock "refresh_cache-043ac602-6c4a-4275-b019-5d717579e7c0" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.008778] env[62600]: DEBUG nova.network.neutron [req-5bb2717e-15fe-4d56-9d33-05d6e8d22b1c req-6f681a1c-f14f-48a5-8cae-c088d7f5e61f service nova] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Refreshing network info cache for port a0340049-6048-4c62-8468-1af1b2551a37 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 909.013862] env[62600]: DEBUG oslo_vmware.api [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222693, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.144574] env[62600]: DEBUG nova.compute.utils [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 909.149785] env[62600]: DEBUG nova.compute.manager [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 909.149785] env[62600]: DEBUG nova.network.neutron [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 909.169966] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "88cf521d-f4a3-493a-894d-c7f799dc0443" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.169966] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "88cf521d-f4a3-493a-894d-c7f799dc0443" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.211919] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222689, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.213762] env[62600]: DEBUG nova.policy [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd6b8f15bad243a298b0de9f5f040c44', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ce517935e6a4387ad8385b1fb3658fa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 909.266104] env[62600]: DEBUG oslo_vmware.api [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222690, 'name': PowerOffVM_Task, 'duration_secs': 0.255808} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.266504] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 909.266722] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 909.267037] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a3c7c3cb-74a2-44f8-be61-5a2963d3e75c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.348902] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 909.349315] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 909.349599] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Deleting the datastore file [datastore2] d2e43ba2-4ccf-4625-91ba-78a6f2632461 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 909.349968] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94f2035e-7e05-446b-8585-b24c46495d9c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.358032] env[62600]: DEBUG oslo_vmware.api [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 909.358032] env[62600]: value = "task-1222695" [ 909.358032] env[62600]: _type = "Task" [ 909.358032] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.366808] env[62600]: DEBUG oslo_vmware.api [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222695, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.490037] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': task-1222691, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.503051] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222692, 'name': CreateVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.508555] env[62600]: DEBUG oslo_vmware.api [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222693, 'name': PowerOffVM_Task, 'duration_secs': 0.251209} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.508823] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 909.508997] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 909.509271] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4bc8a7d1-7998-4795-ac2d-d3be8dfb8065 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.541558] env[62600]: DEBUG nova.network.neutron [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Successfully created port: 07d47e0c-0331-4404-9aa9-3568372c18ff {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 909.634611] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 909.634859] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 909.635067] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Deleting the datastore file [datastore2] 48bbbdaf-496d-4780-b467-6b393257535d {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 909.635342] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e255930c-81cb-472e-8444-26681af8b1e9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.643097] env[62600]: DEBUG oslo_vmware.api [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 909.643097] env[62600]: value = "task-1222697" [ 909.643097] env[62600]: _type = "Task" [ 909.643097] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.651743] env[62600]: DEBUG nova.compute.manager [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 909.654381] env[62600]: DEBUG oslo_vmware.api [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222697, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.675540] env[62600]: DEBUG nova.compute.manager [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 909.711878] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222689, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.653027} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.715776] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] d627a701-77cd-4a1d-9e52-0fc4adc93391/d627a701-77cd-4a1d-9e52-0fc4adc93391.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 909.716210] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 909.717019] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6f9c7c72-5821-4e42-9261-8c1769349cc9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.727277] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 909.727277] env[62600]: value = "task-1222698" [ 909.727277] env[62600]: _type = "Task" [ 909.727277] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.736439] env[62600]: DEBUG nova.network.neutron [req-5bb2717e-15fe-4d56-9d33-05d6e8d22b1c req-6f681a1c-f14f-48a5-8cae-c088d7f5e61f service nova] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Updated VIF entry in instance network info cache for port a0340049-6048-4c62-8468-1af1b2551a37. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 909.737039] env[62600]: DEBUG nova.network.neutron [req-5bb2717e-15fe-4d56-9d33-05d6e8d22b1c req-6f681a1c-f14f-48a5-8cae-c088d7f5e61f service nova] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Updating instance_info_cache with network_info: [{"id": "a0340049-6048-4c62-8468-1af1b2551a37", "address": "fa:16:3e:32:5b:6a", "network": {"id": "5ef67e8a-5613-4cdb-9f35-0a18720657ff", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1644754500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bab74e3ae78248909dbcd483abbb8da2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0340049-60", "ovs_interfaceid": "a0340049-6048-4c62-8468-1af1b2551a37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.748383] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222698, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.871945] env[62600]: DEBUG oslo_vmware.api [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222695, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.188851} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.872311] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 909.872493] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 909.872678] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 909.872877] env[62600]: INFO nova.compute.manager [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Took 1.14 seconds to destroy the instance on the hypervisor. [ 909.873238] env[62600]: DEBUG oslo.service.loopingcall [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 909.873453] env[62600]: DEBUG nova.compute.manager [-] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 909.873453] env[62600]: DEBUG nova.network.neutron [-] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 909.993596] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': task-1222691, 'name': ReconfigVM_Task, 'duration_secs': 0.603152} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.996891] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Reconfigured VM instance instance-0000004c to attach disk [datastore1] 4eaa22b9-5834-4425-82a9-d16e316f3a52/4eaa22b9-5834-4425-82a9-d16e316f3a52.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 910.000307] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6948f3ea-f8e9-4ee9-8ba1-7c4f9fff71c7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.008752] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222692, 'name': CreateVM_Task, 'duration_secs': 0.527717} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.010020] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 910.010407] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Waiting for the task: (returnval){ [ 910.010407] env[62600]: value = "task-1222699" [ 910.010407] env[62600]: _type = "Task" [ 910.010407] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.011053] env[62600]: DEBUG oslo_concurrency.lockutils [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.011233] env[62600]: DEBUG oslo_concurrency.lockutils [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.012295] env[62600]: DEBUG oslo_concurrency.lockutils [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 910.012295] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b205836c-176f-4dae-8fac-2427e7cc8d90 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.020641] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-024ed6a5-3a01-499c-959f-9a61f6227a5b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.027620] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': task-1222699, 'name': Rename_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.030891] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 910.030891] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]527e43f4-f4c7-2a2e-400e-9963c91cdaf5" [ 910.030891] env[62600]: _type = "Task" [ 910.030891] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.035331] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79da7fe5-1aca-4d09-8575-457eca91ca42 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.041631] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]527e43f4-f4c7-2a2e-400e-9963c91cdaf5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.070161] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c05ec40-4e86-4003-bf5d-9fa1780b0eb5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.079318] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e66ca93-939f-4363-a7da-89c58c276e67 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.096840] env[62600]: DEBUG nova.compute.provider_tree [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 910.153208] env[62600]: DEBUG oslo_vmware.api [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222697, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.247552} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.154284] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 910.155337] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 910.155337] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 910.155337] env[62600]: INFO nova.compute.manager [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Took 1.21 seconds to destroy the instance on the hypervisor. [ 910.155337] env[62600]: DEBUG oslo.service.loopingcall [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.155337] env[62600]: DEBUG nova.compute.manager [-] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 910.155507] env[62600]: DEBUG nova.network.neutron [-] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 910.195500] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.236433] env[62600]: DEBUG nova.compute.manager [req-1b0b73e0-8e38-4281-a14a-3b83d0d0883a req-40147893-fef9-4d9b-85b0-150f993b31fd service nova] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Received event network-vif-deleted-e04d7519-77e4-4593-b7ec-5fc549cfce8f {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 910.236651] env[62600]: INFO nova.compute.manager [req-1b0b73e0-8e38-4281-a14a-3b83d0d0883a req-40147893-fef9-4d9b-85b0-150f993b31fd service nova] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Neutron deleted interface e04d7519-77e4-4593-b7ec-5fc549cfce8f; detaching it from the instance and deleting it from the info cache [ 910.236834] env[62600]: DEBUG nova.network.neutron [req-1b0b73e0-8e38-4281-a14a-3b83d0d0883a req-40147893-fef9-4d9b-85b0-150f993b31fd service nova] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.239353] env[62600]: DEBUG oslo_concurrency.lockutils [req-5bb2717e-15fe-4d56-9d33-05d6e8d22b1c req-6f681a1c-f14f-48a5-8cae-c088d7f5e61f service nova] Releasing lock "refresh_cache-043ac602-6c4a-4275-b019-5d717579e7c0" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.244476] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222698, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.192072} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.244730] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 910.245557] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3cef045-dd52-47b0-922b-e91336c3662d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.269747] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] d627a701-77cd-4a1d-9e52-0fc4adc93391/d627a701-77cd-4a1d-9e52-0fc4adc93391.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 910.270302] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a68d04f-59f2-4fb7-8bef-7c8f70b1a92e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.291404] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 910.291404] env[62600]: value = "task-1222700" [ 910.291404] env[62600]: _type = "Task" [ 910.291404] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.304175] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222700, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.522110] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': task-1222699, 'name': Rename_Task, 'duration_secs': 0.160905} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.522470] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 910.522799] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eea89b80-8103-4ddc-9ad4-bed85f9dd0af {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.530149] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Waiting for the task: (returnval){ [ 910.530149] env[62600]: value = "task-1222701" [ 910.530149] env[62600]: _type = "Task" [ 910.530149] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.541704] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': task-1222701, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.544932] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]527e43f4-f4c7-2a2e-400e-9963c91cdaf5, 'name': SearchDatastore_Task, 'duration_secs': 0.014551} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.545246] env[62600]: DEBUG oslo_concurrency.lockutils [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.545489] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 910.545728] env[62600]: DEBUG oslo_concurrency.lockutils [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.545992] env[62600]: DEBUG oslo_concurrency.lockutils [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.546326] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 910.546326] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1242ff66-07be-4120-a89a-17bb85233a0d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.555696] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 910.555882] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 910.556608] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7157a020-2782-474f-b98c-69ec150aec57 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.562383] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 910.562383] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525d13ec-e3e2-9e40-c17a-f0f771e3da85" [ 910.562383] env[62600]: _type = "Task" [ 910.562383] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.570954] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525d13ec-e3e2-9e40-c17a-f0f771e3da85, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.619074] env[62600]: ERROR nova.scheduler.client.report [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [req-cbd8df91-1da2-4e80-af41-1c185f3dfd67] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 664af347-7147-4bf5-9019-9ae15cb4aa82. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-cbd8df91-1da2-4e80-af41-1c185f3dfd67"}]} [ 910.635046] env[62600]: DEBUG nova.scheduler.client.report [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Refreshing inventories for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 910.648796] env[62600]: DEBUG nova.scheduler.client.report [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Updating ProviderTree inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 910.649067] env[62600]: DEBUG nova.compute.provider_tree [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 910.662058] env[62600]: DEBUG nova.scheduler.client.report [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Refreshing aggregate associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, aggregates: None {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 910.665101] env[62600]: DEBUG nova.compute.manager [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 910.671353] env[62600]: DEBUG nova.network.neutron [-] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.682270] env[62600]: DEBUG nova.scheduler.client.report [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Refreshing trait associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 910.697018] env[62600]: DEBUG nova.virt.hardware [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 910.697533] env[62600]: DEBUG nova.virt.hardware [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 910.697533] env[62600]: DEBUG nova.virt.hardware [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 910.697757] env[62600]: DEBUG nova.virt.hardware [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 910.698147] env[62600]: DEBUG nova.virt.hardware [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 910.698147] env[62600]: DEBUG nova.virt.hardware [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 910.698311] env[62600]: DEBUG nova.virt.hardware [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 910.699435] env[62600]: DEBUG nova.virt.hardware [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 910.699435] env[62600]: DEBUG nova.virt.hardware [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 910.699435] env[62600]: DEBUG nova.virt.hardware [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 910.699435] env[62600]: DEBUG nova.virt.hardware [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 910.699970] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d99ad1cc-c2c1-44fb-a23e-e8cf3ca14887 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.712202] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb94e032-0094-45a3-9117-f841715dbe99 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.739822] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-40ae11a8-da9e-49de-80c2-6fd5bd3fc2f9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.750257] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba44201c-5a36-4e25-8c9e-713e850f092e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.781871] env[62600]: DEBUG nova.compute.manager [req-1b0b73e0-8e38-4281-a14a-3b83d0d0883a req-40147893-fef9-4d9b-85b0-150f993b31fd service nova] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Detach interface failed, port_id=e04d7519-77e4-4593-b7ec-5fc549cfce8f, reason: Instance d2e43ba2-4ccf-4625-91ba-78a6f2632461 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 910.803965] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222700, 'name': ReconfigVM_Task, 'duration_secs': 0.2896} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.804307] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Reconfigured VM instance instance-00000049 to attach disk [datastore1] d627a701-77cd-4a1d-9e52-0fc4adc93391/d627a701-77cd-4a1d-9e52-0fc4adc93391.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 910.804981] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cfa10270-c271-458c-8bf4-0ee047af6da1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.812758] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 910.812758] env[62600]: value = "task-1222702" [ 910.812758] env[62600]: _type = "Task" [ 910.812758] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.822823] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222702, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.864060] env[62600]: DEBUG nova.network.neutron [-] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.927252] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a32d640-b9cc-4e88-b52d-2a8cf2ec88ac {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.936817] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7b59346-7b2f-4fe1-b452-9c83e5b7a974 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.970535] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54c7291a-5059-4e38-9c00-dbf5d64aa01b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.978860] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a524151a-06f3-4611-92e0-28009811ea48 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.992689] env[62600]: DEBUG nova.compute.provider_tree [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 911.044126] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': task-1222701, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.072949] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525d13ec-e3e2-9e40-c17a-f0f771e3da85, 'name': SearchDatastore_Task, 'duration_secs': 0.019799} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.073765] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d28250a4-5365-49d1-85fd-088d86e10a93 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.079357] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 911.079357] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526e08b2-51f3-3d2c-0b86-9c0d3e1c2051" [ 911.079357] env[62600]: _type = "Task" [ 911.079357] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.087588] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526e08b2-51f3-3d2c-0b86-9c0d3e1c2051, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.173810] env[62600]: INFO nova.compute.manager [-] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Took 1.30 seconds to deallocate network for instance. [ 911.325080] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222702, 'name': Rename_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.367616] env[62600]: INFO nova.compute.manager [-] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Took 1.21 seconds to deallocate network for instance. [ 911.384301] env[62600]: DEBUG nova.compute.manager [req-3c8df75d-01ee-4bff-8642-54040a74e5e1 req-b39663a4-f4c5-4349-8443-dbf0718b682c service nova] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Received event network-vif-plugged-07d47e0c-0331-4404-9aa9-3568372c18ff {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.384669] env[62600]: DEBUG oslo_concurrency.lockutils [req-3c8df75d-01ee-4bff-8642-54040a74e5e1 req-b39663a4-f4c5-4349-8443-dbf0718b682c service nova] Acquiring lock "9ed08410-a4cb-43c4-9271-2c04d9c87eac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.384930] env[62600]: DEBUG oslo_concurrency.lockutils [req-3c8df75d-01ee-4bff-8642-54040a74e5e1 req-b39663a4-f4c5-4349-8443-dbf0718b682c service nova] Lock "9ed08410-a4cb-43c4-9271-2c04d9c87eac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.385158] env[62600]: DEBUG oslo_concurrency.lockutils [req-3c8df75d-01ee-4bff-8642-54040a74e5e1 req-b39663a4-f4c5-4349-8443-dbf0718b682c service nova] Lock "9ed08410-a4cb-43c4-9271-2c04d9c87eac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.385320] env[62600]: DEBUG nova.compute.manager [req-3c8df75d-01ee-4bff-8642-54040a74e5e1 req-b39663a4-f4c5-4349-8443-dbf0718b682c service nova] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] No waiting events found dispatching network-vif-plugged-07d47e0c-0331-4404-9aa9-3568372c18ff {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 911.385515] env[62600]: WARNING nova.compute.manager [req-3c8df75d-01ee-4bff-8642-54040a74e5e1 req-b39663a4-f4c5-4349-8443-dbf0718b682c service nova] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Received unexpected event network-vif-plugged-07d47e0c-0331-4404-9aa9-3568372c18ff for instance with vm_state building and task_state spawning. [ 911.529322] env[62600]: DEBUG nova.scheduler.client.report [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 103 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 911.529322] env[62600]: DEBUG nova.compute.provider_tree [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 103 to 104 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 911.529322] env[62600]: DEBUG nova.compute.provider_tree [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 911.545372] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': task-1222701, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.589803] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526e08b2-51f3-3d2c-0b86-9c0d3e1c2051, 'name': SearchDatastore_Task, 'duration_secs': 0.009833} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.590111] env[62600]: DEBUG oslo_concurrency.lockutils [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.590389] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 043ac602-6c4a-4275-b019-5d717579e7c0/043ac602-6c4a-4275-b019-5d717579e7c0.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 911.590645] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ae4cd3c7-54ae-44ea-860b-a71f52b40b61 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.600041] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 911.600041] env[62600]: value = "task-1222703" [ 911.600041] env[62600]: _type = "Task" [ 911.600041] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.606794] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222703, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.680489] env[62600]: DEBUG oslo_concurrency.lockutils [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.824333] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222702, 'name': Rename_Task, 'duration_secs': 0.83999} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.824759] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 911.825013] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fe2c3be9-a231-449f-b39c-0603274c0b01 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.831995] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 911.831995] env[62600]: value = "task-1222704" [ 911.831995] env[62600]: _type = "Task" [ 911.831995] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.841909] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222704, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.877567] env[62600]: DEBUG oslo_concurrency.lockutils [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.899218] env[62600]: DEBUG nova.network.neutron [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Successfully updated port: 07d47e0c-0331-4404-9aa9-3568372c18ff {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 911.928444] env[62600]: DEBUG nova.compute.manager [req-29a170a0-7dd0-4cf5-91cf-8e4e1f831846 req-40578247-8bc7-41ea-9507-c8e18c3e6a48 service nova] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Received event network-changed-07d47e0c-0331-4404-9aa9-3568372c18ff {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.928691] env[62600]: DEBUG nova.compute.manager [req-29a170a0-7dd0-4cf5-91cf-8e4e1f831846 req-40578247-8bc7-41ea-9507-c8e18c3e6a48 service nova] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Refreshing instance network info cache due to event network-changed-07d47e0c-0331-4404-9aa9-3568372c18ff. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 911.928804] env[62600]: DEBUG oslo_concurrency.lockutils [req-29a170a0-7dd0-4cf5-91cf-8e4e1f831846 req-40578247-8bc7-41ea-9507-c8e18c3e6a48 service nova] Acquiring lock "refresh_cache-9ed08410-a4cb-43c4-9271-2c04d9c87eac" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.928951] env[62600]: DEBUG oslo_concurrency.lockutils [req-29a170a0-7dd0-4cf5-91cf-8e4e1f831846 req-40578247-8bc7-41ea-9507-c8e18c3e6a48 service nova] Acquired lock "refresh_cache-9ed08410-a4cb-43c4-9271-2c04d9c87eac" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.929132] env[62600]: DEBUG nova.network.neutron [req-29a170a0-7dd0-4cf5-91cf-8e4e1f831846 req-40578247-8bc7-41ea-9507-c8e18c3e6a48 service nova] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Refreshing network info cache for port 07d47e0c-0331-4404-9aa9-3568372c18ff {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 912.034609] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.398s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.035218] env[62600]: DEBUG nova.compute.manager [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 912.038115] env[62600]: DEBUG oslo_concurrency.lockutils [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.084s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.039660] env[62600]: INFO nova.compute.claims [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 912.054044] env[62600]: DEBUG oslo_vmware.api [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': task-1222701, 'name': PowerOnVM_Task, 'duration_secs': 1.034265} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.054346] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 912.054601] env[62600]: INFO nova.compute.manager [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Took 9.22 seconds to spawn the instance on the hypervisor. [ 912.054897] env[62600]: DEBUG nova.compute.manager [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 912.055705] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e73459a-ccec-4076-8c4b-4ad4e306236c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.109826] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222703, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509159} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.109826] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 043ac602-6c4a-4275-b019-5d717579e7c0/043ac602-6c4a-4275-b019-5d717579e7c0.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 912.109826] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 912.109826] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b844e222-4403-43bd-8078-93903091daf3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.117719] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 912.117719] env[62600]: value = "task-1222705" [ 912.117719] env[62600]: _type = "Task" [ 912.117719] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.126060] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222705, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.323431] env[62600]: DEBUG nova.compute.manager [req-525f97ea-5529-46a4-8b38-82cba7c2221c req-36eafa2f-6cb3-4cdf-a2ee-b0c27f2d0637 service nova] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Received event network-vif-deleted-f01caa0e-4aff-41e9-a7ed-0511436b131d {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 912.343042] env[62600]: DEBUG oslo_vmware.api [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222704, 'name': PowerOnVM_Task, 'duration_secs': 0.501157} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.343185] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 912.343409] env[62600]: DEBUG nova.compute.manager [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 912.344213] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd5d9ab-b85a-4685-bb0a-fc34808a8271 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.402029] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Acquiring lock "refresh_cache-9ed08410-a4cb-43c4-9271-2c04d9c87eac" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.464863] env[62600]: DEBUG nova.network.neutron [req-29a170a0-7dd0-4cf5-91cf-8e4e1f831846 req-40578247-8bc7-41ea-9507-c8e18c3e6a48 service nova] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 912.549137] env[62600]: DEBUG nova.compute.utils [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 912.550466] env[62600]: DEBUG nova.compute.manager [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 912.550633] env[62600]: DEBUG nova.network.neutron [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 912.573308] env[62600]: INFO nova.compute.manager [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Took 37.49 seconds to build instance. [ 912.612580] env[62600]: DEBUG nova.policy [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3e59998217a4b18a6f3f01142a5e440', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f91091f83ee4a2091507ca994e3d52f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 912.615742] env[62600]: DEBUG nova.network.neutron [req-29a170a0-7dd0-4cf5-91cf-8e4e1f831846 req-40578247-8bc7-41ea-9507-c8e18c3e6a48 service nova] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.630613] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222705, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063977} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.631028] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 912.632330] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccf97ace-acba-487c-9741-d97db1d69b62 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.660055] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 043ac602-6c4a-4275-b019-5d717579e7c0/043ac602-6c4a-4275-b019-5d717579e7c0.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 912.660473] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d54dd642-19c1-4d80-ba2f-b4612c55724f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.683084] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 912.683084] env[62600]: value = "task-1222706" [ 912.683084] env[62600]: _type = "Task" [ 912.683084] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.692902] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222706, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.868157] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.977490] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Acquiring lock "4eaa22b9-5834-4425-82a9-d16e316f3a52" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.042112] env[62600]: DEBUG nova.network.neutron [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Successfully created port: 11fe3228-11ce-46c7-8bf9-5376ed0ce904 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 913.053855] env[62600]: DEBUG nova.compute.manager [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 913.075838] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6926f332-cac3-45fb-9935-35ff4768805b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Lock "4eaa22b9-5834-4425-82a9-d16e316f3a52" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.995s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.076131] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Lock "4eaa22b9-5834-4425-82a9-d16e316f3a52" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.099s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.076630] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Acquiring lock "4eaa22b9-5834-4425-82a9-d16e316f3a52-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.076692] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Lock "4eaa22b9-5834-4425-82a9-d16e316f3a52-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.076832] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Lock "4eaa22b9-5834-4425-82a9-d16e316f3a52-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.081379] env[62600]: INFO nova.compute.manager [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Terminating instance [ 913.083457] env[62600]: DEBUG nova.compute.manager [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 913.083666] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 913.084545] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb70407d-bc68-4a5d-9b5e-ed6e4271789b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.092179] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 913.093082] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9cc956c6-b4c4-4541-bdc8-cf3c179104df {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.100738] env[62600]: DEBUG oslo_vmware.api [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Waiting for the task: (returnval){ [ 913.100738] env[62600]: value = "task-1222707" [ 913.100738] env[62600]: _type = "Task" [ 913.100738] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.109754] env[62600]: DEBUG oslo_vmware.api [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': task-1222707, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.118724] env[62600]: DEBUG oslo_concurrency.lockutils [req-29a170a0-7dd0-4cf5-91cf-8e4e1f831846 req-40578247-8bc7-41ea-9507-c8e18c3e6a48 service nova] Releasing lock "refresh_cache-9ed08410-a4cb-43c4-9271-2c04d9c87eac" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.119076] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Acquired lock "refresh_cache-9ed08410-a4cb-43c4-9271-2c04d9c87eac" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.119240] env[62600]: DEBUG nova.network.neutron [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 913.194908] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222706, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.332351] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887d05db-2cc0-4d97-9d54-7a346643dd06 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.339918] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b6cb285-35ae-48ad-aa26-6ae6b344ed32 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.371823] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07412f93-b626-4c8c-80ec-746d25b21b2b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.379248] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c429e9f-b9ec-45ba-a328-914624dfce62 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.392219] env[62600]: DEBUG nova.compute.provider_tree [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.611695] env[62600]: DEBUG oslo_vmware.api [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': task-1222707, 'name': PowerOffVM_Task, 'duration_secs': 0.216957} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.612116] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 913.612367] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 913.612678] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-154ceb97-60eb-4a95-b73f-0f65d85e72a2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.676018] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 913.676018] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 913.676018] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Deleting the datastore file [datastore1] 4eaa22b9-5834-4425-82a9-d16e316f3a52 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 913.676018] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ca6a31e2-9ca3-4cdb-a25f-bf8e15353881 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.685212] env[62600]: DEBUG oslo_vmware.api [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Waiting for the task: (returnval){ [ 913.685212] env[62600]: value = "task-1222709" [ 913.685212] env[62600]: _type = "Task" [ 913.685212] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.696425] env[62600]: DEBUG nova.network.neutron [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 913.701761] env[62600]: DEBUG oslo_vmware.api [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': task-1222709, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.705612] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222706, 'name': ReconfigVM_Task, 'duration_secs': 0.803653} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.705882] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 043ac602-6c4a-4275-b019-5d717579e7c0/043ac602-6c4a-4275-b019-5d717579e7c0.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 913.706561] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-919d06b4-d16d-4a53-8430-e5aa1e11b274 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.713500] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 913.713500] env[62600]: value = "task-1222710" [ 913.713500] env[62600]: _type = "Task" [ 913.713500] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.722416] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222710, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.895747] env[62600]: DEBUG nova.scheduler.client.report [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 913.962179] env[62600]: DEBUG nova.network.neutron [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Updating instance_info_cache with network_info: [{"id": "07d47e0c-0331-4404-9aa9-3568372c18ff", "address": "fa:16:3e:7a:c4:1b", "network": {"id": "3d319aac-cfd8-4c41-9552-bf4e99a26eb7", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-997575530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ce517935e6a4387ad8385b1fb3658fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e614f8e-6b11-4b6b-a421-904bca6acd91", "external-id": "nsx-vlan-transportzone-923", "segmentation_id": 923, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07d47e0c-03", "ovs_interfaceid": "07d47e0c-0331-4404-9aa9-3568372c18ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.065954] env[62600]: DEBUG nova.compute.manager [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 914.095834] env[62600]: DEBUG nova.virt.hardware [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 914.095953] env[62600]: DEBUG nova.virt.hardware [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 914.096071] env[62600]: DEBUG nova.virt.hardware [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 914.096268] env[62600]: DEBUG nova.virt.hardware [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 914.096422] env[62600]: DEBUG nova.virt.hardware [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 914.096575] env[62600]: DEBUG nova.virt.hardware [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 914.096787] env[62600]: DEBUG nova.virt.hardware [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 914.096951] env[62600]: DEBUG nova.virt.hardware [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 914.097140] env[62600]: DEBUG nova.virt.hardware [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 914.097365] env[62600]: DEBUG nova.virt.hardware [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 914.097485] env[62600]: DEBUG nova.virt.hardware [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 914.098786] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84001882-94c6-4090-ac1a-e6c06f7a382d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.110143] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3ac8749-5d54-4d16-b71d-f22b13f25077 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.192773] env[62600]: DEBUG oslo_vmware.api [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Task: {'id': task-1222709, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139614} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.193070] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 914.193267] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 914.193451] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 914.193631] env[62600]: INFO nova.compute.manager [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Took 1.11 seconds to destroy the instance on the hypervisor. [ 914.193878] env[62600]: DEBUG oslo.service.loopingcall [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.194113] env[62600]: DEBUG nova.compute.manager [-] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 914.194265] env[62600]: DEBUG nova.network.neutron [-] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 914.222705] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222710, 'name': Rename_Task, 'duration_secs': 0.156917} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.222975] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 914.223227] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0a893598-7281-41fe-a1b1-5e8a1ee1324d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.229220] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 914.229220] env[62600]: value = "task-1222711" [ 914.229220] env[62600]: _type = "Task" [ 914.229220] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.236400] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222711, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.403252] env[62600]: DEBUG oslo_concurrency.lockutils [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.365s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.403803] env[62600]: DEBUG nova.compute.manager [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 914.407078] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.501s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.466431] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Releasing lock "refresh_cache-9ed08410-a4cb-43c4-9271-2c04d9c87eac" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.467190] env[62600]: DEBUG nova.compute.manager [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Instance network_info: |[{"id": "07d47e0c-0331-4404-9aa9-3568372c18ff", "address": "fa:16:3e:7a:c4:1b", "network": {"id": "3d319aac-cfd8-4c41-9552-bf4e99a26eb7", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-997575530-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2ce517935e6a4387ad8385b1fb3658fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2e614f8e-6b11-4b6b-a421-904bca6acd91", "external-id": "nsx-vlan-transportzone-923", "segmentation_id": 923, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap07d47e0c-03", "ovs_interfaceid": "07d47e0c-0331-4404-9aa9-3568372c18ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 914.468364] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7a:c4:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2e614f8e-6b11-4b6b-a421-904bca6acd91', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '07d47e0c-0331-4404-9aa9-3568372c18ff', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 914.476340] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Creating folder: Project (2ce517935e6a4387ad8385b1fb3658fa). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 914.476648] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-66379a89-8fb5-462f-95b7-8d18014eb597 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.489914] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Created folder: Project (2ce517935e6a4387ad8385b1fb3658fa) in parent group-v264198. [ 914.490123] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Creating folder: Instances. Parent ref: group-v264325. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 914.490375] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-105f582b-8e00-4d61-a3b5-c86d97ba989d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.500943] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Created folder: Instances in parent group-v264325. [ 914.501091] env[62600]: DEBUG oslo.service.loopingcall [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.501293] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 914.501499] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-63a7a78c-0b54-44a4-a8a2-7eb9941e66ba {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.523634] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 914.523634] env[62600]: value = "task-1222714" [ 914.523634] env[62600]: _type = "Task" [ 914.523634] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.541637] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222714, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.557882] env[62600]: DEBUG nova.compute.manager [req-03f9f1e3-dcaa-40b9-a588-212966e28de4 req-99834f4e-a60d-4343-a42c-405f23ecad03 service nova] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Received event network-vif-plugged-11fe3228-11ce-46c7-8bf9-5376ed0ce904 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.558218] env[62600]: DEBUG oslo_concurrency.lockutils [req-03f9f1e3-dcaa-40b9-a588-212966e28de4 req-99834f4e-a60d-4343-a42c-405f23ecad03 service nova] Acquiring lock "6447a26d-fc30-41eb-bd42-94341a86c2e8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.558524] env[62600]: DEBUG oslo_concurrency.lockutils [req-03f9f1e3-dcaa-40b9-a588-212966e28de4 req-99834f4e-a60d-4343-a42c-405f23ecad03 service nova] Lock "6447a26d-fc30-41eb-bd42-94341a86c2e8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.558765] env[62600]: DEBUG oslo_concurrency.lockutils [req-03f9f1e3-dcaa-40b9-a588-212966e28de4 req-99834f4e-a60d-4343-a42c-405f23ecad03 service nova] Lock "6447a26d-fc30-41eb-bd42-94341a86c2e8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.558960] env[62600]: DEBUG nova.compute.manager [req-03f9f1e3-dcaa-40b9-a588-212966e28de4 req-99834f4e-a60d-4343-a42c-405f23ecad03 service nova] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] No waiting events found dispatching network-vif-plugged-11fe3228-11ce-46c7-8bf9-5376ed0ce904 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 914.560071] env[62600]: WARNING nova.compute.manager [req-03f9f1e3-dcaa-40b9-a588-212966e28de4 req-99834f4e-a60d-4343-a42c-405f23ecad03 service nova] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Received unexpected event network-vif-plugged-11fe3228-11ce-46c7-8bf9-5376ed0ce904 for instance with vm_state building and task_state spawning. [ 914.604321] env[62600]: DEBUG nova.network.neutron [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Successfully updated port: 11fe3228-11ce-46c7-8bf9-5376ed0ce904 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 914.606603] env[62600]: DEBUG oslo_concurrency.lockutils [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "d627a701-77cd-4a1d-9e52-0fc4adc93391" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.607119] env[62600]: DEBUG oslo_concurrency.lockutils [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "d627a701-77cd-4a1d-9e52-0fc4adc93391" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.607533] env[62600]: DEBUG oslo_concurrency.lockutils [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "d627a701-77cd-4a1d-9e52-0fc4adc93391-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.607820] env[62600]: DEBUG oslo_concurrency.lockutils [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "d627a701-77cd-4a1d-9e52-0fc4adc93391-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.608215] env[62600]: DEBUG oslo_concurrency.lockutils [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "d627a701-77cd-4a1d-9e52-0fc4adc93391-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.611013] env[62600]: INFO nova.compute.manager [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Terminating instance [ 914.614319] env[62600]: DEBUG nova.compute.manager [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 914.614648] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 914.616627] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48095d48-071b-4e53-bb20-ce9e261de32a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.626618] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 914.626909] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b41b40ca-0e83-4c18-b867-aa5b2eaf1a69 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.633497] env[62600]: DEBUG oslo_vmware.api [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 914.633497] env[62600]: value = "task-1222715" [ 914.633497] env[62600]: _type = "Task" [ 914.633497] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.643321] env[62600]: DEBUG nova.compute.manager [req-910db9ca-3a33-4ee5-83c7-3d0fc69a35ae req-0cab7eb6-f137-422c-bef4-ec5566a57bd3 service nova] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Received event network-vif-deleted-abd09e35-cd67-49d8-960d-1713a9907a35 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.643321] env[62600]: INFO nova.compute.manager [req-910db9ca-3a33-4ee5-83c7-3d0fc69a35ae req-0cab7eb6-f137-422c-bef4-ec5566a57bd3 service nova] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Neutron deleted interface abd09e35-cd67-49d8-960d-1713a9907a35; detaching it from the instance and deleting it from the info cache [ 914.643321] env[62600]: DEBUG nova.network.neutron [req-910db9ca-3a33-4ee5-83c7-3d0fc69a35ae req-0cab7eb6-f137-422c-bef4-ec5566a57bd3 service nova] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.646871] env[62600]: DEBUG oslo_vmware.api [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222715, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.742719] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222711, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.911072] env[62600]: DEBUG nova.compute.utils [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 914.920956] env[62600]: DEBUG nova.compute.manager [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 914.921283] env[62600]: DEBUG nova.network.neutron [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 914.972815] env[62600]: DEBUG nova.policy [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a2416189869c4cbe950d6fe3d4310eb8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '48e1772c9a6c4671ab89df2e8b40cae1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 915.037022] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222714, 'name': CreateVM_Task, 'duration_secs': 0.308826} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.037216] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 915.037977] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.038183] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.038524] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 915.038827] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5c231cd-cd86-4d85-8273-6465686c8edc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.042339] env[62600]: DEBUG nova.network.neutron [-] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.045432] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for the task: (returnval){ [ 915.045432] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52745646-819a-c4cc-24b2-a12e991dff53" [ 915.045432] env[62600]: _type = "Task" [ 915.045432] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.055385] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52745646-819a-c4cc-24b2-a12e991dff53, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.107324] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "refresh_cache-6447a26d-fc30-41eb-bd42-94341a86c2e8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.107488] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "refresh_cache-6447a26d-fc30-41eb-bd42-94341a86c2e8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.107645] env[62600]: DEBUG nova.network.neutron [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 915.146815] env[62600]: DEBUG oslo_vmware.api [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222715, 'name': PowerOffVM_Task, 'duration_secs': 0.196542} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.147124] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 915.147307] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 915.148028] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a4b2da8c-984a-4347-a6c5-8c3146565465 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.149763] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f171c268-5453-4f49-bb82-4c6844839a47 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.158251] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b4e5628-d763-4d9c-a500-cf2129ad664c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.188474] env[62600]: DEBUG nova.compute.manager [req-910db9ca-3a33-4ee5-83c7-3d0fc69a35ae req-0cab7eb6-f137-422c-bef4-ec5566a57bd3 service nova] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Detach interface failed, port_id=abd09e35-cd67-49d8-960d-1713a9907a35, reason: Instance 4eaa22b9-5834-4425-82a9-d16e316f3a52 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 915.192445] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "f18df958-2885-4d8c-a9ad-140faaed1178" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.192669] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "f18df958-2885-4d8c-a9ad-140faaed1178" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.236061] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 915.236299] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 915.236496] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleting the datastore file [datastore1] d627a701-77cd-4a1d-9e52-0fc4adc93391 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 915.236813] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-debf22b3-be06-4eb8-a3ff-ffc09d9f2ddc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.241255] env[62600]: DEBUG oslo_vmware.api [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222711, 'name': PowerOnVM_Task, 'duration_secs': 0.798771} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.241809] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 915.242071] env[62600]: INFO nova.compute.manager [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Took 8.95 seconds to spawn the instance on the hypervisor. [ 915.242257] env[62600]: DEBUG nova.compute.manager [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 915.242988] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecec80ba-1c10-4aef-8320-4929911d3374 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.246279] env[62600]: DEBUG oslo_vmware.api [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 915.246279] env[62600]: value = "task-1222717" [ 915.246279] env[62600]: _type = "Task" [ 915.246279] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.258974] env[62600]: DEBUG oslo_vmware.api [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222717, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.284358] env[62600]: DEBUG nova.network.neutron [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Successfully created port: 32f5d56d-a029-4469-ad3c-77a4ca51bc02 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 915.421861] env[62600]: DEBUG nova.compute.manager [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 915.447327] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance afe838cc-e086-4986-87ec-4e1266bcaf60 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 915.447553] env[62600]: WARNING nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance dde6df10-618a-40a8-b33f-efc0ca3a9287 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 915.447707] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 9d899d96-9f4f-41d1-a368-3fde5efc110a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 915.447850] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance b967fb11-e70e-4e17-b769-38da581bd83b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 915.447969] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance d66c404d-8fd1-4fb7-a3b9-f21854c7e735 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 915.448131] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 1b856aff-66be-4fa5-b8b7-124ac24fb4d1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 915.448876] env[62600]: WARNING nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 915.448876] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance ab11cdd1-80a7-43c4-afac-24c7930238c1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 915.448876] env[62600]: WARNING nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance f6957b5a-5da8-4205-92e5-2f08bf948e88 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 915.448876] env[62600]: WARNING nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 107ed783-10a3-4f07-a809-a5a076c2b904 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 915.448876] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance d627a701-77cd-4a1d-9e52-0fc4adc93391 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 915.449087] env[62600]: WARNING nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance d2e43ba2-4ccf-4625-91ba-78a6f2632461 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 915.449221] env[62600]: WARNING nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 48bbbdaf-496d-4780-b467-6b393257535d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 915.449361] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 4eaa22b9-5834-4425-82a9-d16e316f3a52 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 915.449480] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 043ac602-6c4a-4275-b019-5d717579e7c0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 915.449595] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 9ed08410-a4cb-43c4-9271-2c04d9c87eac actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 915.449726] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 6447a26d-fc30-41eb-bd42-94341a86c2e8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 915.449859] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 248a5371-6ff7-4da7-ae97-d638bea0123a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 915.548083] env[62600]: INFO nova.compute.manager [-] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Took 1.35 seconds to deallocate network for instance. [ 915.561303] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52745646-819a-c4cc-24b2-a12e991dff53, 'name': SearchDatastore_Task, 'duration_secs': 0.01007} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.561303] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.561303] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 915.561481] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.561621] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.561798] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 915.562089] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e68020b0-2745-4675-8fd7-d4415e22273e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.569896] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 915.570117] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 915.570825] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1265d663-91bc-4f54-a3bd-846c42486ad8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.575895] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for the task: (returnval){ [ 915.575895] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d5b364-e81c-ff3f-c228-ba49be1ba9cc" [ 915.575895] env[62600]: _type = "Task" [ 915.575895] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.583439] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d5b364-e81c-ff3f-c228-ba49be1ba9cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.664072] env[62600]: DEBUG nova.network.neutron [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 915.694902] env[62600]: DEBUG nova.compute.manager [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 915.766733] env[62600]: DEBUG oslo_vmware.api [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222717, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.118307} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.767302] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 915.767496] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 915.767682] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 915.767858] env[62600]: INFO nova.compute.manager [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Took 1.15 seconds to destroy the instance on the hypervisor. [ 915.768536] env[62600]: DEBUG oslo.service.loopingcall [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.768977] env[62600]: INFO nova.compute.manager [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Took 35.85 seconds to build instance. [ 915.770168] env[62600]: DEBUG nova.compute.manager [-] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 915.770273] env[62600]: DEBUG nova.network.neutron [-] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 915.911052] env[62600]: DEBUG nova.network.neutron [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Updating instance_info_cache with network_info: [{"id": "11fe3228-11ce-46c7-8bf9-5376ed0ce904", "address": "fa:16:3e:e3:f6:79", "network": {"id": "1d2052cf-5be1-48e7-887a-18f367b7823f", "bridge": "br-int", "label": "tempest-ServersTestJSON-106099679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f91091f83ee4a2091507ca994e3d52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11fe3228-11", "ovs_interfaceid": "11fe3228-11ce-46c7-8bf9-5376ed0ce904", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.953187] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance c2f38d56-72fa-4bf0-a2cb-13d666571466 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 916.057178] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.088175] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d5b364-e81c-ff3f-c228-ba49be1ba9cc, 'name': SearchDatastore_Task, 'duration_secs': 0.007834} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.088553] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-838aa0bc-92a4-4e65-9d4d-58641c24ab03 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.094107] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for the task: (returnval){ [ 916.094107] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5224c701-63ff-1835-2f54-6be4ddd4bd1e" [ 916.094107] env[62600]: _type = "Task" [ 916.094107] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.102281] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5224c701-63ff-1835-2f54-6be4ddd4bd1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.217924] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.271218] env[62600]: DEBUG oslo_concurrency.lockutils [None req-712a64b5-57fa-4ea7-9179-f6da2e1070a9 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "043ac602-6c4a-4275-b019-5d717579e7c0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.360s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.414657] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "refresh_cache-6447a26d-fc30-41eb-bd42-94341a86c2e8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.414989] env[62600]: DEBUG nova.compute.manager [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Instance network_info: |[{"id": "11fe3228-11ce-46c7-8bf9-5376ed0ce904", "address": "fa:16:3e:e3:f6:79", "network": {"id": "1d2052cf-5be1-48e7-887a-18f367b7823f", "bridge": "br-int", "label": "tempest-ServersTestJSON-106099679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f91091f83ee4a2091507ca994e3d52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11fe3228-11", "ovs_interfaceid": "11fe3228-11ce-46c7-8bf9-5376ed0ce904", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 916.415459] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e3:f6:79', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '028bae2d-fe6c-4207-b4a3-3fab45fbf1d6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '11fe3228-11ce-46c7-8bf9-5376ed0ce904', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 916.423231] env[62600]: DEBUG oslo.service.loopingcall [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 916.423532] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 916.423655] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c949d1ae-a04c-4e63-82ae-231abf74af62 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.441742] env[62600]: DEBUG nova.compute.manager [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 916.449436] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 916.449436] env[62600]: value = "task-1222718" [ 916.449436] env[62600]: _type = "Task" [ 916.449436] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.457921] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 88cf521d-f4a3-493a-894d-c7f799dc0443 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 916.459178] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222718, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.461209] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22bcaf77-daa0-4478-836b-8608b1e31959 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "043ac602-6c4a-4275-b019-5d717579e7c0" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.461447] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22bcaf77-daa0-4478-836b-8608b1e31959 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "043ac602-6c4a-4275-b019-5d717579e7c0" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.461625] env[62600]: DEBUG nova.compute.manager [None req-22bcaf77-daa0-4478-836b-8608b1e31959 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 916.462775] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baa10a41-2dd3-41f1-9457-eba62e0ab4a9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.468427] env[62600]: DEBUG nova.virt.hardware [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 916.468921] env[62600]: DEBUG nova.virt.hardware [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 916.468921] env[62600]: DEBUG nova.virt.hardware [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 916.469852] env[62600]: DEBUG nova.virt.hardware [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 916.469852] env[62600]: DEBUG nova.virt.hardware [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 916.469852] env[62600]: DEBUG nova.virt.hardware [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 916.469852] env[62600]: DEBUG nova.virt.hardware [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 916.469852] env[62600]: DEBUG nova.virt.hardware [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 916.469852] env[62600]: DEBUG nova.virt.hardware [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 916.471248] env[62600]: DEBUG nova.virt.hardware [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 916.471248] env[62600]: DEBUG nova.virt.hardware [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 916.471248] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d758c29-7fd2-414f-8ab7-0eecb4b61002 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.475687] env[62600]: DEBUG nova.compute.manager [None req-22bcaf77-daa0-4478-836b-8608b1e31959 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62600) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 916.476264] env[62600]: DEBUG nova.objects.instance [None req-22bcaf77-daa0-4478-836b-8608b1e31959 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lazy-loading 'flavor' on Instance uuid 043ac602-6c4a-4275-b019-5d717579e7c0 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.483894] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22bb066d-3673-4525-b8dc-087d43a2593e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.529104] env[62600]: DEBUG nova.network.neutron [-] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.586326] env[62600]: DEBUG nova.compute.manager [req-62091bef-d2ed-4260-bc01-c2d15fbcfa41 req-09d61ec3-6d04-4bba-be35-09c638e3148f service nova] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Received event network-changed-11fe3228-11ce-46c7-8bf9-5376ed0ce904 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.586562] env[62600]: DEBUG nova.compute.manager [req-62091bef-d2ed-4260-bc01-c2d15fbcfa41 req-09d61ec3-6d04-4bba-be35-09c638e3148f service nova] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Refreshing instance network info cache due to event network-changed-11fe3228-11ce-46c7-8bf9-5376ed0ce904. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 916.586798] env[62600]: DEBUG oslo_concurrency.lockutils [req-62091bef-d2ed-4260-bc01-c2d15fbcfa41 req-09d61ec3-6d04-4bba-be35-09c638e3148f service nova] Acquiring lock "refresh_cache-6447a26d-fc30-41eb-bd42-94341a86c2e8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.586953] env[62600]: DEBUG oslo_concurrency.lockutils [req-62091bef-d2ed-4260-bc01-c2d15fbcfa41 req-09d61ec3-6d04-4bba-be35-09c638e3148f service nova] Acquired lock "refresh_cache-6447a26d-fc30-41eb-bd42-94341a86c2e8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.587296] env[62600]: DEBUG nova.network.neutron [req-62091bef-d2ed-4260-bc01-c2d15fbcfa41 req-09d61ec3-6d04-4bba-be35-09c638e3148f service nova] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Refreshing network info cache for port 11fe3228-11ce-46c7-8bf9-5376ed0ce904 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 916.606720] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5224c701-63ff-1835-2f54-6be4ddd4bd1e, 'name': SearchDatastore_Task, 'duration_secs': 0.009355} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.607012] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.607410] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 9ed08410-a4cb-43c4-9271-2c04d9c87eac/9ed08410-a4cb-43c4-9271-2c04d9c87eac.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 916.607558] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-af5526c3-c752-4580-bc3a-bfe52f407369 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.615459] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for the task: (returnval){ [ 916.615459] env[62600]: value = "task-1222719" [ 916.615459] env[62600]: _type = "Task" [ 916.615459] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.624106] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222719, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.781676] env[62600]: DEBUG nova.network.neutron [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Successfully updated port: 32f5d56d-a029-4469-ad3c-77a4ca51bc02 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 916.960882] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222718, 'name': CreateVM_Task, 'duration_secs': 0.387739} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.961093] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 916.961825] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.962030] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.962381] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 916.963121] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance f18df958-2885-4d8c-a9ad-140faaed1178 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 916.963445] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 916.963522] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2816MB phys_disk=200GB used_disk=12GB total_vcpus=48 used_vcpus=12 pci_stats=[] {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 916.965964] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23b0f9b9-b5cf-495c-8d51-1338b3d7939c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.971671] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 916.971671] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5249c875-91f0-4c05-330c-aa35235d36d7" [ 916.971671] env[62600]: _type = "Task" [ 916.971671] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.982689] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5249c875-91f0-4c05-330c-aa35235d36d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.986830] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-22bcaf77-daa0-4478-836b-8608b1e31959 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 916.987407] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e4e8daf-fb96-4214-a5ca-2be98db8e196 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.995139] env[62600]: DEBUG oslo_vmware.api [None req-22bcaf77-daa0-4478-836b-8608b1e31959 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 916.995139] env[62600]: value = "task-1222720" [ 916.995139] env[62600]: _type = "Task" [ 916.995139] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.009663] env[62600]: DEBUG oslo_vmware.api [None req-22bcaf77-daa0-4478-836b-8608b1e31959 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222720, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.031811] env[62600]: INFO nova.compute.manager [-] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Took 1.26 seconds to deallocate network for instance. [ 917.127078] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222719, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475232} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.130045] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 9ed08410-a4cb-43c4-9271-2c04d9c87eac/9ed08410-a4cb-43c4-9271-2c04d9c87eac.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 917.130045] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 917.130401] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d59b4889-1744-4edd-8cbe-62d7ba460ea4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.139826] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for the task: (returnval){ [ 917.139826] env[62600]: value = "task-1222721" [ 917.139826] env[62600]: _type = "Task" [ 917.139826] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.147689] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222721, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.279848] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-811e0804-04c3-4be3-a2bb-bd6be3c90543 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.290111] env[62600]: DEBUG oslo_concurrency.lockutils [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.290307] env[62600]: DEBUG oslo_concurrency.lockutils [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquired lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.290455] env[62600]: DEBUG nova.network.neutron [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 917.292423] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50549f3c-cbfe-46a8-a797-55a8fbe53f24 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.326253] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef1b66e5-43d1-4ad5-9cf8-ecac57ef3013 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.332156] env[62600]: DEBUG nova.network.neutron [req-62091bef-d2ed-4260-bc01-c2d15fbcfa41 req-09d61ec3-6d04-4bba-be35-09c638e3148f service nova] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Updated VIF entry in instance network info cache for port 11fe3228-11ce-46c7-8bf9-5376ed0ce904. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 917.332156] env[62600]: DEBUG nova.network.neutron [req-62091bef-d2ed-4260-bc01-c2d15fbcfa41 req-09d61ec3-6d04-4bba-be35-09c638e3148f service nova] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Updating instance_info_cache with network_info: [{"id": "11fe3228-11ce-46c7-8bf9-5376ed0ce904", "address": "fa:16:3e:e3:f6:79", "network": {"id": "1d2052cf-5be1-48e7-887a-18f367b7823f", "bridge": "br-int", "label": "tempest-ServersTestJSON-106099679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f91091f83ee4a2091507ca994e3d52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11fe3228-11", "ovs_interfaceid": "11fe3228-11ce-46c7-8bf9-5376ed0ce904", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.338402] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a5eda9-ffb7-4ff0-8ba3-bf423b0c94be {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.353189] env[62600]: DEBUG nova.compute.provider_tree [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 917.482244] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5249c875-91f0-4c05-330c-aa35235d36d7, 'name': SearchDatastore_Task, 'duration_secs': 0.023093} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.482599] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.482799] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 917.483066] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.483245] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.483435] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 917.483697] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a1404e00-f177-4b99-9205-2c679e32c3a7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.491383] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 917.491564] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 917.492277] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fd0ea86-f056-4d1d-8e36-0c690d841b57 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.502164] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 917.502164] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525d80e5-a09b-edbb-3bf4-50599e64c10a" [ 917.502164] env[62600]: _type = "Task" [ 917.502164] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.505089] env[62600]: DEBUG oslo_vmware.api [None req-22bcaf77-daa0-4478-836b-8608b1e31959 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222720, 'name': PowerOffVM_Task, 'duration_secs': 0.187019} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.508063] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-22bcaf77-daa0-4478-836b-8608b1e31959 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 917.508277] env[62600]: DEBUG nova.compute.manager [None req-22bcaf77-daa0-4478-836b-8608b1e31959 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 917.508970] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68cd4554-b013-45ff-8055-cda04e098ee6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.515893] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525d80e5-a09b-edbb-3bf4-50599e64c10a, 'name': SearchDatastore_Task, 'duration_secs': 0.00913} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.519515] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4eff3a57-d15d-4875-a027-f454d7e30ca5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.524460] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 917.524460] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52eab1e5-5fb8-72a8-f2e8-1c8eba888141" [ 917.524460] env[62600]: _type = "Task" [ 917.524460] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.531687] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52eab1e5-5fb8-72a8-f2e8-1c8eba888141, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.538647] env[62600]: DEBUG oslo_concurrency.lockutils [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.649407] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222721, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.823758] env[62600]: DEBUG nova.network.neutron [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 917.833341] env[62600]: DEBUG oslo_concurrency.lockutils [req-62091bef-d2ed-4260-bc01-c2d15fbcfa41 req-09d61ec3-6d04-4bba-be35-09c638e3148f service nova] Releasing lock "refresh_cache-6447a26d-fc30-41eb-bd42-94341a86c2e8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.833620] env[62600]: DEBUG nova.compute.manager [req-62091bef-d2ed-4260-bc01-c2d15fbcfa41 req-09d61ec3-6d04-4bba-be35-09c638e3148f service nova] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Received event network-vif-deleted-92bc4652-839c-47b3-bcf5-2cc0c43d5537 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.874453] env[62600]: ERROR nova.scheduler.client.report [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [req-83c92b46-c47c-476d-aee9-5e813d3fea29] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 664af347-7147-4bf5-9019-9ae15cb4aa82. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-83c92b46-c47c-476d-aee9-5e813d3fea29"}]} [ 917.889963] env[62600]: DEBUG nova.scheduler.client.report [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Refreshing inventories for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 917.903833] env[62600]: DEBUG nova.scheduler.client.report [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Updating ProviderTree inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 917.904062] env[62600]: DEBUG nova.compute.provider_tree [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 917.916910] env[62600]: DEBUG nova.scheduler.client.report [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Refreshing aggregate associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, aggregates: None {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 917.934813] env[62600]: DEBUG nova.scheduler.client.report [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Refreshing trait associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 917.957967] env[62600]: DEBUG nova.network.neutron [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Updating instance_info_cache with network_info: [{"id": "32f5d56d-a029-4469-ad3c-77a4ca51bc02", "address": "fa:16:3e:da:8c:98", "network": {"id": "036dc114-6ea0-46cb-a958-4a6137e7e85f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-728627542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e1772c9a6c4671ab89df2e8b40cae1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32f5d56d-a0", "ovs_interfaceid": "32f5d56d-a029-4469-ad3c-77a4ca51bc02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.024240] env[62600]: DEBUG oslo_concurrency.lockutils [None req-22bcaf77-daa0-4478-836b-8608b1e31959 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "043ac602-6c4a-4275-b019-5d717579e7c0" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.563s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.036697] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52eab1e5-5fb8-72a8-f2e8-1c8eba888141, 'name': SearchDatastore_Task, 'duration_secs': 0.007816} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.037237] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.037536] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 6447a26d-fc30-41eb-bd42-94341a86c2e8/6447a26d-fc30-41eb-bd42-94341a86c2e8.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 918.037802] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c490290f-a95c-46be-95b2-f5a1c9b45305 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.044618] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 918.044618] env[62600]: value = "task-1222722" [ 918.044618] env[62600]: _type = "Task" [ 918.044618] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.054629] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222722, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.157064] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222721, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.201532] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f74ef97-e96a-4c48-b05b-d32104395ec7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.208635] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b22ab96-89d8-4188-b7ea-d56199510ff5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.240574] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac237f29-5ebe-4638-bdb1-ea9284d6aded {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.248743] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d126a215-ca63-4495-a21d-b046ef664feb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.262922] env[62600]: DEBUG nova.compute.provider_tree [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 918.460954] env[62600]: DEBUG oslo_concurrency.lockutils [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Releasing lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.461416] env[62600]: DEBUG nova.compute.manager [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Instance network_info: |[{"id": "32f5d56d-a029-4469-ad3c-77a4ca51bc02", "address": "fa:16:3e:da:8c:98", "network": {"id": "036dc114-6ea0-46cb-a958-4a6137e7e85f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-728627542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e1772c9a6c4671ab89df2e8b40cae1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32f5d56d-a0", "ovs_interfaceid": "32f5d56d-a029-4469-ad3c-77a4ca51bc02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 918.462320] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:da:8c:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3e05affa-2640-435e-a124-0ee8a6ab1152', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '32f5d56d-a029-4469-ad3c-77a4ca51bc02', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 918.470044] env[62600]: DEBUG oslo.service.loopingcall [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 918.470340] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 918.470581] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-efbfc8b6-ed42-431e-9339-2a151e6d74cf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.490646] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 918.490646] env[62600]: value = "task-1222723" [ 918.490646] env[62600]: _type = "Task" [ 918.490646] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.498322] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222723, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.554810] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222722, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457874} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.554941] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 6447a26d-fc30-41eb-bd42-94341a86c2e8/6447a26d-fc30-41eb-bd42-94341a86c2e8.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 918.555237] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 918.555542] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a85b846-e11f-4649-92ce-aa3217a915cc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.561614] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 918.561614] env[62600]: value = "task-1222724" [ 918.561614] env[62600]: _type = "Task" [ 918.561614] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.569265] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222724, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.650677] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222721, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.023681} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.650953] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 918.651755] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-610d4d18-83a8-4f88-bc21-2af12c474a14 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.677953] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 9ed08410-a4cb-43c4-9271-2c04d9c87eac/9ed08410-a4cb-43c4-9271-2c04d9c87eac.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 918.678339] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-890724bb-f105-4eb7-88db-4892f5e2b65f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.698132] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for the task: (returnval){ [ 918.698132] env[62600]: value = "task-1222725" [ 918.698132] env[62600]: _type = "Task" [ 918.698132] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.706070] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222725, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.792989] env[62600]: DEBUG nova.scheduler.client.report [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 105 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 918.793375] env[62600]: DEBUG nova.compute.provider_tree [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 105 to 106 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 918.793578] env[62600]: DEBUG nova.compute.provider_tree [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 918.813760] env[62600]: DEBUG nova.compute.manager [req-edfd5ced-2125-49a6-887a-d26bed1d9148 req-2f04c00e-b5d4-4850-afdb-05b11aa03b3b service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Received event network-vif-plugged-32f5d56d-a029-4469-ad3c-77a4ca51bc02 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.814046] env[62600]: DEBUG oslo_concurrency.lockutils [req-edfd5ced-2125-49a6-887a-d26bed1d9148 req-2f04c00e-b5d4-4850-afdb-05b11aa03b3b service nova] Acquiring lock "248a5371-6ff7-4da7-ae97-d638bea0123a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.814442] env[62600]: DEBUG oslo_concurrency.lockutils [req-edfd5ced-2125-49a6-887a-d26bed1d9148 req-2f04c00e-b5d4-4850-afdb-05b11aa03b3b service nova] Lock "248a5371-6ff7-4da7-ae97-d638bea0123a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.814585] env[62600]: DEBUG oslo_concurrency.lockutils [req-edfd5ced-2125-49a6-887a-d26bed1d9148 req-2f04c00e-b5d4-4850-afdb-05b11aa03b3b service nova] Lock "248a5371-6ff7-4da7-ae97-d638bea0123a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.814787] env[62600]: DEBUG nova.compute.manager [req-edfd5ced-2125-49a6-887a-d26bed1d9148 req-2f04c00e-b5d4-4850-afdb-05b11aa03b3b service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] No waiting events found dispatching network-vif-plugged-32f5d56d-a029-4469-ad3c-77a4ca51bc02 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 918.814979] env[62600]: WARNING nova.compute.manager [req-edfd5ced-2125-49a6-887a-d26bed1d9148 req-2f04c00e-b5d4-4850-afdb-05b11aa03b3b service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Received unexpected event network-vif-plugged-32f5d56d-a029-4469-ad3c-77a4ca51bc02 for instance with vm_state building and task_state spawning. [ 918.815163] env[62600]: DEBUG nova.compute.manager [req-edfd5ced-2125-49a6-887a-d26bed1d9148 req-2f04c00e-b5d4-4850-afdb-05b11aa03b3b service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Received event network-changed-32f5d56d-a029-4469-ad3c-77a4ca51bc02 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.815333] env[62600]: DEBUG nova.compute.manager [req-edfd5ced-2125-49a6-887a-d26bed1d9148 req-2f04c00e-b5d4-4850-afdb-05b11aa03b3b service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Refreshing instance network info cache due to event network-changed-32f5d56d-a029-4469-ad3c-77a4ca51bc02. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 918.815564] env[62600]: DEBUG oslo_concurrency.lockutils [req-edfd5ced-2125-49a6-887a-d26bed1d9148 req-2f04c00e-b5d4-4850-afdb-05b11aa03b3b service nova] Acquiring lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.815710] env[62600]: DEBUG oslo_concurrency.lockutils [req-edfd5ced-2125-49a6-887a-d26bed1d9148 req-2f04c00e-b5d4-4850-afdb-05b11aa03b3b service nova] Acquired lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.815873] env[62600]: DEBUG nova.network.neutron [req-edfd5ced-2125-49a6-887a-d26bed1d9148 req-2f04c00e-b5d4-4850-afdb-05b11aa03b3b service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Refreshing network info cache for port 32f5d56d-a029-4469-ad3c-77a4ca51bc02 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 919.001058] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222723, 'name': CreateVM_Task, 'duration_secs': 0.36934} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.001283] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 919.001926] env[62600]: DEBUG oslo_concurrency.lockutils [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.002132] env[62600]: DEBUG oslo_concurrency.lockutils [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.002459] env[62600]: DEBUG oslo_concurrency.lockutils [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 919.002715] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53416066-16a1-43e5-8152-e92431d2c803 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.007082] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 919.007082] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c4c116-cd6e-e647-dac1-fbc886b9dde7" [ 919.007082] env[62600]: _type = "Task" [ 919.007082] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.016561] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c4c116-cd6e-e647-dac1-fbc886b9dde7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.070758] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222724, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061921} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.071000] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 919.071793] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-044ffc07-3011-4e77-a8f0-01e647d2baea {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.093486] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 6447a26d-fc30-41eb-bd42-94341a86c2e8/6447a26d-fc30-41eb-bd42-94341a86c2e8.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 919.093486] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6448804a-5e46-47b7-ac22-34998b282af6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.112921] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 919.112921] env[62600]: value = "task-1222726" [ 919.112921] env[62600]: _type = "Task" [ 919.112921] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.121808] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222726, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.207913] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222725, 'name': ReconfigVM_Task, 'duration_secs': 0.342043} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.209624] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 9ed08410-a4cb-43c4-9271-2c04d9c87eac/9ed08410-a4cb-43c4-9271-2c04d9c87eac.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 919.209624] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c7ec4fc4-4b7a-465d-8ac3-5fbc7bfc1432 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.215675] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for the task: (returnval){ [ 919.215675] env[62600]: value = "task-1222727" [ 919.215675] env[62600]: _type = "Task" [ 919.215675] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.223909] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222727, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.298988] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62600) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 919.299306] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.892s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.299628] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.818s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.299843] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.302164] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.566s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.302402] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.304233] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.588s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.304454] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.306219] env[62600]: DEBUG oslo_concurrency.lockutils [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.227s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.307941] env[62600]: INFO nova.compute.claims [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 919.334262] env[62600]: INFO nova.scheduler.client.report [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Deleted allocations for instance dde6df10-618a-40a8-b33f-efc0ca3a9287 [ 919.336553] env[62600]: INFO nova.scheduler.client.report [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Deleted allocations for instance c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f [ 919.348789] env[62600]: INFO nova.scheduler.client.report [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Deleted allocations for instance f6957b5a-5da8-4205-92e5-2f08bf948e88 [ 919.357578] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "043ac602-6c4a-4275-b019-5d717579e7c0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.357674] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "043ac602-6c4a-4275-b019-5d717579e7c0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.357838] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "043ac602-6c4a-4275-b019-5d717579e7c0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.358044] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "043ac602-6c4a-4275-b019-5d717579e7c0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.358225] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "043ac602-6c4a-4275-b019-5d717579e7c0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.360059] env[62600]: INFO nova.compute.manager [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Terminating instance [ 919.362134] env[62600]: DEBUG nova.compute.manager [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 919.362350] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 919.363498] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9581f1bb-a4f3-4503-a453-b7324cbaccce {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.370861] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 919.371191] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-36c72eb8-9213-4a73-8172-c4786fcd5877 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.446175] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 919.446175] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 919.446175] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Deleting the datastore file [datastore1] 043ac602-6c4a-4275-b019-5d717579e7c0 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 919.446346] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2bc2efb2-c127-4b80-8955-82614679ac7f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.452169] env[62600]: DEBUG oslo_vmware.api [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 919.452169] env[62600]: value = "task-1222729" [ 919.452169] env[62600]: _type = "Task" [ 919.452169] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.460634] env[62600]: DEBUG oslo_vmware.api [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222729, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.517315] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c4c116-cd6e-e647-dac1-fbc886b9dde7, 'name': SearchDatastore_Task, 'duration_secs': 0.009046} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.519681] env[62600]: DEBUG oslo_concurrency.lockutils [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.519934] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 919.520219] env[62600]: DEBUG oslo_concurrency.lockutils [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.520377] env[62600]: DEBUG oslo_concurrency.lockutils [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.520559] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 919.520828] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-52c69019-f3f6-4bbb-9511-7a6320fcf135 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.528337] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 919.528337] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 919.529025] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8a8b654-f2c0-480c-99a6-0ba2c2e6b4c1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.533802] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 919.533802] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52dce718-dd59-0a8d-1148-e03213d3016c" [ 919.533802] env[62600]: _type = "Task" [ 919.533802] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.542297] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52dce718-dd59-0a8d-1148-e03213d3016c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.555777] env[62600]: DEBUG nova.network.neutron [req-edfd5ced-2125-49a6-887a-d26bed1d9148 req-2f04c00e-b5d4-4850-afdb-05b11aa03b3b service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Updated VIF entry in instance network info cache for port 32f5d56d-a029-4469-ad3c-77a4ca51bc02. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 919.556210] env[62600]: DEBUG nova.network.neutron [req-edfd5ced-2125-49a6-887a-d26bed1d9148 req-2f04c00e-b5d4-4850-afdb-05b11aa03b3b service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Updating instance_info_cache with network_info: [{"id": "32f5d56d-a029-4469-ad3c-77a4ca51bc02", "address": "fa:16:3e:da:8c:98", "network": {"id": "036dc114-6ea0-46cb-a958-4a6137e7e85f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-728627542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e1772c9a6c4671ab89df2e8b40cae1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32f5d56d-a0", "ovs_interfaceid": "32f5d56d-a029-4469-ad3c-77a4ca51bc02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.622170] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222726, 'name': ReconfigVM_Task, 'duration_secs': 0.274871} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.622459] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 6447a26d-fc30-41eb-bd42-94341a86c2e8/6447a26d-fc30-41eb-bd42-94341a86c2e8.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 919.623132] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9300f316-1153-4ce6-8fec-0458862bf6e5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.629469] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 919.629469] env[62600]: value = "task-1222730" [ 919.629469] env[62600]: _type = "Task" [ 919.629469] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.637835] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222730, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.726789] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222727, 'name': Rename_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.847916] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84aa184d-da12-48ab-97e2-e2ca46d82154 tempest-AttachVolumeTestJSON-74082861 tempest-AttachVolumeTestJSON-74082861-project-member] Lock "dde6df10-618a-40a8-b33f-efc0ca3a9287" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.322s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.848899] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5a0a058c-4bcd-4658-8fe4-b64259126df0 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.188s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.857122] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a6e79d4c-8cf6-4b22-964a-1e7ea38bdfe0 tempest-ServerAddressesTestJSON-1779077296 tempest-ServerAddressesTestJSON-1779077296-project-member] Lock "f6957b5a-5da8-4205-92e5-2f08bf948e88" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.069s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.961886] env[62600]: DEBUG oslo_vmware.api [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222729, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154828} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.963491] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 919.963491] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 919.963491] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 919.963491] env[62600]: INFO nova.compute.manager [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Took 0.60 seconds to destroy the instance on the hypervisor. [ 919.963491] env[62600]: DEBUG oslo.service.loopingcall [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.963491] env[62600]: DEBUG nova.compute.manager [-] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 919.963491] env[62600]: DEBUG nova.network.neutron [-] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 920.044311] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52dce718-dd59-0a8d-1148-e03213d3016c, 'name': SearchDatastore_Task, 'duration_secs': 0.009738} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.047023] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13788f5c-88b5-4416-b06e-2126fce432cb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.050278] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 920.050278] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52bc2b3c-eb01-6ad3-d9a3-df19e42a211c" [ 920.050278] env[62600]: _type = "Task" [ 920.050278] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.057937] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52bc2b3c-eb01-6ad3-d9a3-df19e42a211c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.058517] env[62600]: DEBUG oslo_concurrency.lockutils [req-edfd5ced-2125-49a6-887a-d26bed1d9148 req-2f04c00e-b5d4-4850-afdb-05b11aa03b3b service nova] Releasing lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.146947] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222730, 'name': Rename_Task, 'duration_secs': 0.132327} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.147272] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 920.147523] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f9f7156c-f1f0-4630-8011-1988728a993d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.153759] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 920.153759] env[62600]: value = "task-1222731" [ 920.153759] env[62600]: _type = "Task" [ 920.153759] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.165326] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222731, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.226357] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222727, 'name': Rename_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.568560] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52bc2b3c-eb01-6ad3-d9a3-df19e42a211c, 'name': SearchDatastore_Task, 'duration_secs': 0.008727} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.568560] env[62600]: DEBUG oslo_concurrency.lockutils [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.568560] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 248a5371-6ff7-4da7-ae97-d638bea0123a/248a5371-6ff7-4da7-ae97-d638bea0123a.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 920.569051] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e6292e3b-e651-435d-b255-7acf686d1882 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.578745] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 920.578745] env[62600]: value = "task-1222732" [ 920.578745] env[62600]: _type = "Task" [ 920.578745] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.586698] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222732, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.595506] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ae7f8c-d062-4542-bf51-8eea6d0b115f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.604055] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba60658-78c4-41b5-a25f-289229d1ea73 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.635252] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d14679e2-52ab-46d0-82a0-95302a54095e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.643756] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a00437b-aba8-440b-a8dc-0039fd8e4054 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.659193] env[62600]: DEBUG nova.compute.provider_tree [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.668941] env[62600]: DEBUG oslo_vmware.api [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222731, 'name': PowerOnVM_Task, 'duration_secs': 0.448057} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.669344] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 920.669419] env[62600]: INFO nova.compute.manager [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Took 6.60 seconds to spawn the instance on the hypervisor. [ 920.669561] env[62600]: DEBUG nova.compute.manager [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 920.670977] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2585be6a-f2a2-4727-82b7-70ce496b7358 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.731332] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222727, 'name': Rename_Task, 'duration_secs': 1.140417} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.731332] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 920.731332] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-449b952e-b038-4b52-8b6f-443acfc78236 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.737823] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for the task: (returnval){ [ 920.737823] env[62600]: value = "task-1222734" [ 920.737823] env[62600]: _type = "Task" [ 920.737823] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.747318] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222734, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.775414] env[62600]: DEBUG nova.network.neutron [-] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.972312] env[62600]: DEBUG oslo_concurrency.lockutils [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "9d899d96-9f4f-41d1-a368-3fde5efc110a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.972696] env[62600]: DEBUG oslo_concurrency.lockutils [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "9d899d96-9f4f-41d1-a368-3fde5efc110a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.972888] env[62600]: DEBUG oslo_concurrency.lockutils [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "9d899d96-9f4f-41d1-a368-3fde5efc110a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.973205] env[62600]: DEBUG oslo_concurrency.lockutils [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "9d899d96-9f4f-41d1-a368-3fde5efc110a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.973432] env[62600]: DEBUG oslo_concurrency.lockutils [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "9d899d96-9f4f-41d1-a368-3fde5efc110a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.976303] env[62600]: INFO nova.compute.manager [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Terminating instance [ 920.978813] env[62600]: DEBUG nova.compute.manager [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 920.979279] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 920.980160] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-207f0cba-4544-407a-b36e-ea49c1149aaf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.989697] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 920.989980] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d2bdc95b-a908-4f15-9fd9-802ee10013be {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.997770] env[62600]: DEBUG oslo_vmware.api [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 920.997770] env[62600]: value = "task-1222735" [ 920.997770] env[62600]: _type = "Task" [ 920.997770] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.010320] env[62600]: DEBUG oslo_vmware.api [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222735, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.090074] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222732, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465915} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.090074] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 248a5371-6ff7-4da7-ae97-d638bea0123a/248a5371-6ff7-4da7-ae97-d638bea0123a.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 921.090074] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 921.090074] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-843f25d0-bd82-40f0-a512-7a8465414034 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.097241] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 921.097241] env[62600]: value = "task-1222736" [ 921.097241] env[62600]: _type = "Task" [ 921.097241] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.105565] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222736, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.165465] env[62600]: DEBUG nova.scheduler.client.report [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 921.174726] env[62600]: DEBUG nova.compute.manager [req-32133609-6587-49de-b8c1-701eb58252ac req-f66a223f-8dbd-4216-ae4d-c27a62ed8b19 service nova] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Received event network-vif-deleted-a0340049-6048-4c62-8468-1af1b2551a37 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.192114] env[62600]: INFO nova.compute.manager [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Took 24.29 seconds to build instance. [ 921.248459] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222734, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.278656] env[62600]: INFO nova.compute.manager [-] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Took 1.32 seconds to deallocate network for instance. [ 921.507563] env[62600]: DEBUG oslo_vmware.api [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222735, 'name': PowerOffVM_Task, 'duration_secs': 0.250518} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.508406] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 921.508753] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 921.509157] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3c0277ec-c5b8-473f-a072-324fe08cb0cf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.608851] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222736, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070634} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.610465] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 921.611313] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228fcd84-0662-499a-acee-b7316183322f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.637225] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 248a5371-6ff7-4da7-ae97-d638bea0123a/248a5371-6ff7-4da7-ae97-d638bea0123a.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 921.637564] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ecd0116-2173-4ea9-bf28-3df0f2aaefd5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.664360] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 921.664360] env[62600]: value = "task-1222738" [ 921.664360] env[62600]: _type = "Task" [ 921.664360] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.677878] env[62600]: DEBUG oslo_concurrency.lockutils [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.371s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.678072] env[62600]: DEBUG nova.compute.manager [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 921.681271] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222738, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.683441] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.488s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.683737] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.686740] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.491s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.688972] env[62600]: INFO nova.compute.claims [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 921.692087] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 921.692387] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 921.692659] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Deleting the datastore file [datastore2] 9d899d96-9f4f-41d1-a368-3fde5efc110a {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.694677] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2b5be35f-5373-47f9-bb3a-0122a7366a41 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.703687] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f33b3171-13e4-4bc2-875f-29a02e170c20 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "6447a26d-fc30-41eb-bd42-94341a86c2e8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.261s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.711320] env[62600]: DEBUG oslo_vmware.api [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 921.711320] env[62600]: value = "task-1222739" [ 921.711320] env[62600]: _type = "Task" [ 921.711320] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.719149] env[62600]: DEBUG oslo_vmware.api [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222739, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.727272] env[62600]: INFO nova.scheduler.client.report [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Deleted allocations for instance 107ed783-10a3-4f07-a809-a5a076c2b904 [ 921.749042] env[62600]: DEBUG oslo_vmware.api [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222734, 'name': PowerOnVM_Task, 'duration_secs': 0.529731} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.749571] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 921.749853] env[62600]: INFO nova.compute.manager [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Took 11.08 seconds to spawn the instance on the hypervisor. [ 921.750076] env[62600]: DEBUG nova.compute.manager [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 921.751708] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fefca824-1a72-4c7c-a4ea-f24537daecb7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.789462] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.823616] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "84dc284d-c7f4-4fc2-b539-8d4820429b02" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.823721] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "84dc284d-c7f4-4fc2-b539-8d4820429b02" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.177190] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.186370] env[62600]: DEBUG nova.compute.utils [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 922.186370] env[62600]: DEBUG nova.compute.manager [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 922.186540] env[62600]: DEBUG nova.network.neutron [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 922.207611] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "6447a26d-fc30-41eb-bd42-94341a86c2e8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.208045] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "6447a26d-fc30-41eb-bd42-94341a86c2e8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.208175] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "6447a26d-fc30-41eb-bd42-94341a86c2e8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.208406] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "6447a26d-fc30-41eb-bd42-94341a86c2e8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.208625] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "6447a26d-fc30-41eb-bd42-94341a86c2e8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.214430] env[62600]: INFO nova.compute.manager [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Terminating instance [ 922.224564] env[62600]: DEBUG nova.compute.manager [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 922.224796] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 922.225686] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-331a744b-b2e2-455f-b1a7-16abcf7f478f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.234435] env[62600]: DEBUG oslo_vmware.api [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222739, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145252} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.236657] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 922.236863] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 922.237423] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 922.237423] env[62600]: INFO nova.compute.manager [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Took 1.26 seconds to destroy the instance on the hypervisor. [ 922.237535] env[62600]: DEBUG oslo.service.loopingcall [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.239929] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 922.240781] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d3cf5184-793b-4f38-9517-08002a6e6e25 tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "107ed783-10a3-4f07-a809-a5a076c2b904" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.341s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.241165] env[62600]: DEBUG nova.compute.manager [-] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 922.241269] env[62600]: DEBUG nova.network.neutron [-] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 922.242900] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-56b241f7-6845-4a07-9d46-3a35288c911e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.246831] env[62600]: DEBUG nova.policy [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9bad4b59ce0c4bc1ab9eed7cb40583a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4e088a25c01c40d98d27e22563b4c625', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 922.250908] env[62600]: DEBUG oslo_vmware.api [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 922.250908] env[62600]: value = "task-1222740" [ 922.250908] env[62600]: _type = "Task" [ 922.250908] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.260296] env[62600]: DEBUG oslo_vmware.api [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222740, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.272561] env[62600]: INFO nova.compute.manager [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Took 36.99 seconds to build instance. [ 922.326268] env[62600]: DEBUG nova.compute.manager [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 922.676354] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222738, 'name': ReconfigVM_Task, 'duration_secs': 0.864427} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.676687] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 248a5371-6ff7-4da7-ae97-d638bea0123a/248a5371-6ff7-4da7-ae97-d638bea0123a.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 922.677355] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0373c0f3-78e4-4944-9366-ef129811b397 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.685628] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 922.685628] env[62600]: value = "task-1222742" [ 922.685628] env[62600]: _type = "Task" [ 922.685628] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.689571] env[62600]: DEBUG nova.compute.manager [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 922.697165] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222742, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.758086] env[62600]: DEBUG nova.network.neutron [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Successfully created port: c715a88c-6498-428e-989d-3a1aa2f2e3bd {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 922.771510] env[62600]: DEBUG oslo_vmware.api [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222740, 'name': PowerOffVM_Task, 'duration_secs': 0.304094} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.772331] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 922.772331] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 922.772487] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16d80b05-acc3-4ebd-b7b0-bf99bf1a7bf3 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Lock "9ed08410-a4cb-43c4-9271-2c04d9c87eac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.504s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.772637] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-02549ab4-d411-450d-aa76-5e9ca0b57cde {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.842140] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 922.842296] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 922.842461] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleting the datastore file [datastore1] 6447a26d-fc30-41eb-bd42-94341a86c2e8 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 922.845067] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fb3f16db-6ce9-4bcc-a8f7-158ef0eb531d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.852757] env[62600]: DEBUG oslo_vmware.api [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 922.852757] env[62600]: value = "task-1222745" [ 922.852757] env[62600]: _type = "Task" [ 922.852757] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.861981] env[62600]: DEBUG oslo_vmware.api [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222745, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.865338] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.010865] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e09aee5f-a1cd-443d-b3d1-c53d73dcdda4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.019253] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7074b213-5441-49e4-948a-3f05d59f7da3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.052671] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ca3342a-a8f7-4304-a4bb-e48709106514 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.060611] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-526e9783-8ea5-4d32-9586-1cfb2c1f8766 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.074371] env[62600]: DEBUG nova.compute.provider_tree [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 923.173311] env[62600]: DEBUG nova.network.neutron [-] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.195451] env[62600]: DEBUG nova.compute.manager [req-ed3b98ae-a094-4a58-a467-e5aa5c425252 req-9c7daafd-9975-4baa-8709-80c20ff68c71 service nova] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Received event network-vif-deleted-a3ce62f1-b1fb-40f1-a551-e9d26d299144 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.202356] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222742, 'name': Rename_Task, 'duration_secs': 0.178139} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.202988] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 923.203250] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0b73c652-3782-4a51-8729-80f028c38000 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.209623] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 923.209623] env[62600]: value = "task-1222746" [ 923.209623] env[62600]: _type = "Task" [ 923.209623] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.217313] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222746, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.364112] env[62600]: DEBUG oslo_vmware.api [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222745, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143237} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.364230] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 923.364486] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 923.364696] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 923.364933] env[62600]: INFO nova.compute.manager [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Took 1.14 seconds to destroy the instance on the hypervisor. [ 923.365244] env[62600]: DEBUG oslo.service.loopingcall [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.365482] env[62600]: DEBUG nova.compute.manager [-] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 923.365600] env[62600]: DEBUG nova.network.neutron [-] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 923.599692] env[62600]: ERROR nova.scheduler.client.report [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [req-134b5fc7-bc0c-4ffc-9025-4e6841fd4e10] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 664af347-7147-4bf5-9019-9ae15cb4aa82. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-134b5fc7-bc0c-4ffc-9025-4e6841fd4e10"}]} [ 923.617275] env[62600]: DEBUG nova.scheduler.client.report [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Refreshing inventories for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 923.631659] env[62600]: DEBUG nova.scheduler.client.report [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Updating ProviderTree inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 923.631915] env[62600]: DEBUG nova.compute.provider_tree [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 923.644775] env[62600]: DEBUG nova.scheduler.client.report [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Refreshing aggregate associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, aggregates: None {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 923.664821] env[62600]: DEBUG nova.scheduler.client.report [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Refreshing trait associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 923.675999] env[62600]: INFO nova.compute.manager [-] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Took 1.43 seconds to deallocate network for instance. [ 923.705784] env[62600]: DEBUG nova.compute.manager [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 923.722667] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222746, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.739320] env[62600]: DEBUG nova.virt.hardware [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 923.739658] env[62600]: DEBUG nova.virt.hardware [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 923.739834] env[62600]: DEBUG nova.virt.hardware [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 923.740137] env[62600]: DEBUG nova.virt.hardware [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 923.740376] env[62600]: DEBUG nova.virt.hardware [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 923.740524] env[62600]: DEBUG nova.virt.hardware [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 923.740758] env[62600]: DEBUG nova.virt.hardware [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 923.740943] env[62600]: DEBUG nova.virt.hardware [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 923.741153] env[62600]: DEBUG nova.virt.hardware [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 923.741342] env[62600]: DEBUG nova.virt.hardware [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 923.741537] env[62600]: DEBUG nova.virt.hardware [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 923.742850] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a366c95c-a751-47f5-94da-6d3c4e1a2597 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.757010] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7780789-f6ca-43fb-8cb3-658e4da12c79 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.931873] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4505c8db-a221-4477-a55a-f71626733411 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.940283] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17609eb9-68b4-41ef-94a6-88eb474b43bb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.972980] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beed790b-7837-4542-a41c-a70ad06721f8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.981348] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92c12bbc-be21-4e87-9d51-2f64e278f353 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.994959] env[62600]: DEBUG nova.compute.provider_tree [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 924.092779] env[62600]: DEBUG nova.network.neutron [-] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.140776] env[62600]: DEBUG nova.compute.manager [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.141658] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cab1a6b-a0c2-4767-83d8-5b7b04e191aa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.182863] env[62600]: DEBUG oslo_concurrency.lockutils [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.224380] env[62600]: DEBUG oslo_vmware.api [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222746, 'name': PowerOnVM_Task, 'duration_secs': 0.766387} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.224380] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 924.224380] env[62600]: INFO nova.compute.manager [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Took 7.78 seconds to spawn the instance on the hypervisor. [ 924.224380] env[62600]: DEBUG nova.compute.manager [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.224905] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-024c3220-4db3-430d-9116-121e196376c5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.356570] env[62600]: DEBUG nova.network.neutron [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Successfully updated port: c715a88c-6498-428e-989d-3a1aa2f2e3bd {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 924.525982] env[62600]: DEBUG nova.scheduler.client.report [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 107 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 924.526353] env[62600]: DEBUG nova.compute.provider_tree [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 107 to 108 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 924.526597] env[62600]: DEBUG nova.compute.provider_tree [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 924.595791] env[62600]: INFO nova.compute.manager [-] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Took 1.23 seconds to deallocate network for instance. [ 924.654984] env[62600]: INFO nova.compute.manager [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] instance snapshotting [ 924.657836] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a2d4770-3473-4b78-8524-4e3c729eb554 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.677348] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffcda064-4275-4ddf-9a6c-34cba89b7f04 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.743270] env[62600]: INFO nova.compute.manager [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Took 26.81 seconds to build instance. [ 924.859503] env[62600]: DEBUG oslo_concurrency.lockutils [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Acquiring lock "refresh_cache-c2f38d56-72fa-4bf0-a2cb-13d666571466" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.859673] env[62600]: DEBUG oslo_concurrency.lockutils [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Acquired lock "refresh_cache-c2f38d56-72fa-4bf0-a2cb-13d666571466" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.859829] env[62600]: DEBUG nova.network.neutron [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 925.032350] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.345s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.032866] env[62600]: DEBUG nova.compute.manager [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 925.035741] env[62600]: DEBUG oslo_concurrency.lockutils [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.355s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.035853] env[62600]: DEBUG oslo_concurrency.lockutils [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.039842] env[62600]: DEBUG oslo_concurrency.lockutils [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.160s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.039842] env[62600]: DEBUG oslo_concurrency.lockutils [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.040278] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 12.172s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.040358] env[62600]: DEBUG nova.objects.instance [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62600) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 925.065031] env[62600]: INFO nova.scheduler.client.report [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Deleted allocations for instance 48bbbdaf-496d-4780-b467-6b393257535d [ 925.066765] env[62600]: INFO nova.scheduler.client.report [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Deleted allocations for instance d2e43ba2-4ccf-4625-91ba-78a6f2632461 [ 925.102447] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.188997] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Creating Snapshot of the VM instance {{(pid=62600) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 925.189344] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-db438ef4-39ad-4c3d-b426-7015404b71f1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.200218] env[62600]: DEBUG oslo_vmware.api [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for the task: (returnval){ [ 925.200218] env[62600]: value = "task-1222747" [ 925.200218] env[62600]: _type = "Task" [ 925.200218] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.211677] env[62600]: DEBUG oslo_vmware.api [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222747, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.235026] env[62600]: DEBUG nova.compute.manager [req-89c74122-2fb0-4065-b5d4-2c79a25d2bf0 req-e8cd5025-1a27-4fac-9af9-5d5d08a0296e service nova] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Received event network-vif-deleted-11fe3228-11ce-46c7-8bf9-5376ed0ce904 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.235258] env[62600]: DEBUG nova.compute.manager [req-89c74122-2fb0-4065-b5d4-2c79a25d2bf0 req-e8cd5025-1a27-4fac-9af9-5d5d08a0296e service nova] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Received event network-vif-plugged-c715a88c-6498-428e-989d-3a1aa2f2e3bd {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.235460] env[62600]: DEBUG oslo_concurrency.lockutils [req-89c74122-2fb0-4065-b5d4-2c79a25d2bf0 req-e8cd5025-1a27-4fac-9af9-5d5d08a0296e service nova] Acquiring lock "c2f38d56-72fa-4bf0-a2cb-13d666571466-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.235669] env[62600]: DEBUG oslo_concurrency.lockutils [req-89c74122-2fb0-4065-b5d4-2c79a25d2bf0 req-e8cd5025-1a27-4fac-9af9-5d5d08a0296e service nova] Lock "c2f38d56-72fa-4bf0-a2cb-13d666571466-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.235838] env[62600]: DEBUG oslo_concurrency.lockutils [req-89c74122-2fb0-4065-b5d4-2c79a25d2bf0 req-e8cd5025-1a27-4fac-9af9-5d5d08a0296e service nova] Lock "c2f38d56-72fa-4bf0-a2cb-13d666571466-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.236009] env[62600]: DEBUG nova.compute.manager [req-89c74122-2fb0-4065-b5d4-2c79a25d2bf0 req-e8cd5025-1a27-4fac-9af9-5d5d08a0296e service nova] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] No waiting events found dispatching network-vif-plugged-c715a88c-6498-428e-989d-3a1aa2f2e3bd {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 925.236190] env[62600]: WARNING nova.compute.manager [req-89c74122-2fb0-4065-b5d4-2c79a25d2bf0 req-e8cd5025-1a27-4fac-9af9-5d5d08a0296e service nova] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Received unexpected event network-vif-plugged-c715a88c-6498-428e-989d-3a1aa2f2e3bd for instance with vm_state building and task_state spawning. [ 925.236354] env[62600]: DEBUG nova.compute.manager [req-89c74122-2fb0-4065-b5d4-2c79a25d2bf0 req-e8cd5025-1a27-4fac-9af9-5d5d08a0296e service nova] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Received event network-changed-c715a88c-6498-428e-989d-3a1aa2f2e3bd {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.236513] env[62600]: DEBUG nova.compute.manager [req-89c74122-2fb0-4065-b5d4-2c79a25d2bf0 req-e8cd5025-1a27-4fac-9af9-5d5d08a0296e service nova] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Refreshing instance network info cache due to event network-changed-c715a88c-6498-428e-989d-3a1aa2f2e3bd. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 925.236684] env[62600]: DEBUG oslo_concurrency.lockutils [req-89c74122-2fb0-4065-b5d4-2c79a25d2bf0 req-e8cd5025-1a27-4fac-9af9-5d5d08a0296e service nova] Acquiring lock "refresh_cache-c2f38d56-72fa-4bf0-a2cb-13d666571466" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.244077] env[62600]: DEBUG oslo_concurrency.lockutils [None req-24cafd02-2eac-45d9-8572-8b8c392fbaeb tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "248a5371-6ff7-4da7-ae97-d638bea0123a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.904s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.404887] env[62600]: DEBUG nova.network.neutron [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 925.545394] env[62600]: DEBUG nova.compute.utils [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 925.554023] env[62600]: DEBUG nova.compute.manager [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 925.554023] env[62600]: DEBUG nova.network.neutron [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 925.576678] env[62600]: DEBUG oslo_concurrency.lockutils [None req-081634a7-f1ec-4847-9b17-518a344b089b tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "48bbbdaf-496d-4780-b467-6b393257535d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.652s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.582811] env[62600]: DEBUG oslo_concurrency.lockutils [None req-faa88e37-13ae-4190-bcd0-5212380e2514 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "d2e43ba2-4ccf-4625-91ba-78a6f2632461" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.853s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.652116] env[62600]: DEBUG nova.policy [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fbd1b2fb34d841359ada8fc44bec2986', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b028450f2da445fb83e37adfc86bba68', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 925.675717] env[62600]: DEBUG nova.network.neutron [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Updating instance_info_cache with network_info: [{"id": "c715a88c-6498-428e-989d-3a1aa2f2e3bd", "address": "fa:16:3e:67:c1:e4", "network": {"id": "7ed8725a-d4a1-4d66-8bd3-08c9abf86064", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1314620504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4e088a25c01c40d98d27e22563b4c625", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b98c49ac-0eb7-4311-aa8f-60581b2ce706", "external-id": "nsx-vlan-transportzone-184", "segmentation_id": 184, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc715a88c-64", "ovs_interfaceid": "c715a88c-6498-428e-989d-3a1aa2f2e3bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.712511] env[62600]: DEBUG oslo_vmware.api [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222747, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.054520] env[62600]: DEBUG nova.compute.manager [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 926.058204] env[62600]: DEBUG oslo_concurrency.lockutils [None req-16e41ec4-9baf-4de7-b2b0-5e0b01705714 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.059569] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.003s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.059849] env[62600]: DEBUG nova.objects.instance [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Lazy-loading 'resources' on Instance uuid 4eaa22b9-5834-4425-82a9-d16e316f3a52 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.067845] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquiring lock "afe838cc-e086-4986-87ec-4e1266bcaf60" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.068156] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "afe838cc-e086-4986-87ec-4e1266bcaf60" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.068443] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquiring lock "afe838cc-e086-4986-87ec-4e1266bcaf60-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.068681] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "afe838cc-e086-4986-87ec-4e1266bcaf60-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.068902] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "afe838cc-e086-4986-87ec-4e1266bcaf60-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.072201] env[62600]: INFO nova.compute.manager [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Terminating instance [ 926.073435] env[62600]: DEBUG nova.compute.manager [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 926.073904] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 926.074917] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce2bd262-2d2e-44e4-ad85-31ba12d33cc5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.087156] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 926.088143] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b969b74a-bdd6-4e07-838e-f3b81d2ca097 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.097483] env[62600]: DEBUG oslo_vmware.api [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 926.097483] env[62600]: value = "task-1222748" [ 926.097483] env[62600]: _type = "Task" [ 926.097483] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.108822] env[62600]: DEBUG oslo_vmware.api [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222748, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.178338] env[62600]: DEBUG oslo_concurrency.lockutils [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Releasing lock "refresh_cache-c2f38d56-72fa-4bf0-a2cb-13d666571466" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.178683] env[62600]: DEBUG nova.compute.manager [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Instance network_info: |[{"id": "c715a88c-6498-428e-989d-3a1aa2f2e3bd", "address": "fa:16:3e:67:c1:e4", "network": {"id": "7ed8725a-d4a1-4d66-8bd3-08c9abf86064", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1314620504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4e088a25c01c40d98d27e22563b4c625", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b98c49ac-0eb7-4311-aa8f-60581b2ce706", "external-id": "nsx-vlan-transportzone-184", "segmentation_id": 184, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc715a88c-64", "ovs_interfaceid": "c715a88c-6498-428e-989d-3a1aa2f2e3bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 926.179060] env[62600]: DEBUG oslo_concurrency.lockutils [req-89c74122-2fb0-4065-b5d4-2c79a25d2bf0 req-e8cd5025-1a27-4fac-9af9-5d5d08a0296e service nova] Acquired lock "refresh_cache-c2f38d56-72fa-4bf0-a2cb-13d666571466" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.179252] env[62600]: DEBUG nova.network.neutron [req-89c74122-2fb0-4065-b5d4-2c79a25d2bf0 req-e8cd5025-1a27-4fac-9af9-5d5d08a0296e service nova] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Refreshing network info cache for port c715a88c-6498-428e-989d-3a1aa2f2e3bd {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 926.181185] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:67:c1:e4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b98c49ac-0eb7-4311-aa8f-60581b2ce706', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c715a88c-6498-428e-989d-3a1aa2f2e3bd', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 926.191495] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Creating folder: Project (4e088a25c01c40d98d27e22563b4c625). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 926.192895] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d0424e6b-a89f-48f9-933f-eb4816c0183b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.210021] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Created folder: Project (4e088a25c01c40d98d27e22563b4c625) in parent group-v264198. [ 926.210021] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Creating folder: Instances. Parent ref: group-v264331. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 926.210021] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d5923ac7-b85c-411a-bcdf-3d1df0a4c7ea {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.219284] env[62600]: DEBUG oslo_vmware.api [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222747, 'name': CreateSnapshot_Task, 'duration_secs': 0.761598} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.220844] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Created Snapshot of the VM instance {{(pid=62600) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 926.221871] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0db702c-12db-426a-93d5-02345300081a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.235736] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Created folder: Instances in parent group-v264331. [ 926.236027] env[62600]: DEBUG oslo.service.loopingcall [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.236400] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 926.236665] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-16c057d5-051c-4d1f-90ba-82cde0f5b660 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.258103] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 926.258103] env[62600]: value = "task-1222751" [ 926.258103] env[62600]: _type = "Task" [ 926.258103] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.272365] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222751, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.327792] env[62600]: DEBUG nova.compute.manager [req-fd553666-242a-4c67-a38b-3b50ad501110 req-1cf6ae1f-9c3c-4101-b59a-adaa53cf4747 service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Received event network-changed-32f5d56d-a029-4469-ad3c-77a4ca51bc02 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 926.328014] env[62600]: DEBUG nova.compute.manager [req-fd553666-242a-4c67-a38b-3b50ad501110 req-1cf6ae1f-9c3c-4101-b59a-adaa53cf4747 service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Refreshing instance network info cache due to event network-changed-32f5d56d-a029-4469-ad3c-77a4ca51bc02. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 926.328245] env[62600]: DEBUG oslo_concurrency.lockutils [req-fd553666-242a-4c67-a38b-3b50ad501110 req-1cf6ae1f-9c3c-4101-b59a-adaa53cf4747 service nova] Acquiring lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.328404] env[62600]: DEBUG oslo_concurrency.lockutils [req-fd553666-242a-4c67-a38b-3b50ad501110 req-1cf6ae1f-9c3c-4101-b59a-adaa53cf4747 service nova] Acquired lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.328567] env[62600]: DEBUG nova.network.neutron [req-fd553666-242a-4c67-a38b-3b50ad501110 req-1cf6ae1f-9c3c-4101-b59a-adaa53cf4747 service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Refreshing network info cache for port 32f5d56d-a029-4469-ad3c-77a4ca51bc02 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 926.387224] env[62600]: DEBUG nova.network.neutron [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Successfully created port: 3a8aca44-9791-4493-a1fd-114c9f20fb3f {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 926.619207] env[62600]: DEBUG oslo_vmware.api [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222748, 'name': PowerOffVM_Task, 'duration_secs': 0.215265} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.620459] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 926.620774] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 926.621216] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8848453a-6fa1-4d9f-8a01-a71c9c2cd281 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.689880] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 926.689880] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 926.689880] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Deleting the datastore file [datastore2] afe838cc-e086-4986-87ec-4e1266bcaf60 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 926.689880] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c1ae7754-d2d7-43cc-9678-d83c992e7dcb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.694892] env[62600]: DEBUG oslo_vmware.api [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for the task: (returnval){ [ 926.694892] env[62600]: value = "task-1222753" [ 926.694892] env[62600]: _type = "Task" [ 926.694892] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.706155] env[62600]: DEBUG oslo_vmware.api [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222753, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.742148] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Creating linked-clone VM from snapshot {{(pid=62600) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 926.745652] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ba22822f-b6d5-4976-8c4c-daf20df1cba7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.755496] env[62600]: DEBUG oslo_vmware.api [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for the task: (returnval){ [ 926.755496] env[62600]: value = "task-1222754" [ 926.755496] env[62600]: _type = "Task" [ 926.755496] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.770817] env[62600]: DEBUG oslo_vmware.api [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222754, 'name': CloneVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.775794] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222751, 'name': CreateVM_Task, 'duration_secs': 0.313897} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.776128] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 926.776972] env[62600]: DEBUG oslo_concurrency.lockutils [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.779408] env[62600]: DEBUG oslo_concurrency.lockutils [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.779408] env[62600]: DEBUG oslo_concurrency.lockutils [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 926.779408] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9f0d1f2-b98c-4c64-8b61-f70bcc8f13f6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.788022] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Waiting for the task: (returnval){ [ 926.788022] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52773993-9ff3-3446-6fb7-2ab4cd5a4a9f" [ 926.788022] env[62600]: _type = "Task" [ 926.788022] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.795110] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52773993-9ff3-3446-6fb7-2ab4cd5a4a9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.907355] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8777a227-f4da-47cb-9782-f8f8551ba5df {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.915514] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b3d9f19-c3d2-45da-b1b7-b778f5d16db5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.955187] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf97148-d8c4-4ec3-b9b3-6a59bfb239f2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.967800] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f8d9ea-6b9d-48a6-93c2-5c1b9d431349 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.985460] env[62600]: DEBUG nova.compute.provider_tree [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 927.066314] env[62600]: DEBUG nova.compute.manager [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 927.071125] env[62600]: DEBUG nova.network.neutron [req-89c74122-2fb0-4065-b5d4-2c79a25d2bf0 req-e8cd5025-1a27-4fac-9af9-5d5d08a0296e service nova] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Updated VIF entry in instance network info cache for port c715a88c-6498-428e-989d-3a1aa2f2e3bd. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 927.071478] env[62600]: DEBUG nova.network.neutron [req-89c74122-2fb0-4065-b5d4-2c79a25d2bf0 req-e8cd5025-1a27-4fac-9af9-5d5d08a0296e service nova] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Updating instance_info_cache with network_info: [{"id": "c715a88c-6498-428e-989d-3a1aa2f2e3bd", "address": "fa:16:3e:67:c1:e4", "network": {"id": "7ed8725a-d4a1-4d66-8bd3-08c9abf86064", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1314620504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4e088a25c01c40d98d27e22563b4c625", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b98c49ac-0eb7-4311-aa8f-60581b2ce706", "external-id": "nsx-vlan-transportzone-184", "segmentation_id": 184, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc715a88c-64", "ovs_interfaceid": "c715a88c-6498-428e-989d-3a1aa2f2e3bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.077788] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "784e6d19-36ef-4c01-9e9f-5a083ed08608" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.078039] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "784e6d19-36ef-4c01-9e9f-5a083ed08608" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.101294] env[62600]: DEBUG nova.virt.hardware [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 927.101294] env[62600]: DEBUG nova.virt.hardware [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 927.101294] env[62600]: DEBUG nova.virt.hardware [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 927.101294] env[62600]: DEBUG nova.virt.hardware [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 927.101294] env[62600]: DEBUG nova.virt.hardware [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 927.101294] env[62600]: DEBUG nova.virt.hardware [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 927.101294] env[62600]: DEBUG nova.virt.hardware [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 927.101653] env[62600]: DEBUG nova.virt.hardware [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 927.101653] env[62600]: DEBUG nova.virt.hardware [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 927.101717] env[62600]: DEBUG nova.virt.hardware [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 927.101941] env[62600]: DEBUG nova.virt.hardware [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 927.103592] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-069ddb51-d03b-485a-9171-a383578d632e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.114763] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65bd2e5c-1c9a-4081-887d-d5eb3187942e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.135703] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "81333628-2616-4557-88d1-0e7164e1b16d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.135950] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "81333628-2616-4557-88d1-0e7164e1b16d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.204782] env[62600]: DEBUG oslo_vmware.api [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Task: {'id': task-1222753, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141752} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.205091] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 927.205702] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 927.205702] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 927.205702] env[62600]: INFO nova.compute.manager [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Took 1.13 seconds to destroy the instance on the hypervisor. [ 927.205927] env[62600]: DEBUG oslo.service.loopingcall [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.206130] env[62600]: DEBUG nova.compute.manager [-] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 927.206222] env[62600]: DEBUG nova.network.neutron [-] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 927.214483] env[62600]: DEBUG nova.network.neutron [req-fd553666-242a-4c67-a38b-3b50ad501110 req-1cf6ae1f-9c3c-4101-b59a-adaa53cf4747 service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Updated VIF entry in instance network info cache for port 32f5d56d-a029-4469-ad3c-77a4ca51bc02. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 927.217877] env[62600]: DEBUG nova.network.neutron [req-fd553666-242a-4c67-a38b-3b50ad501110 req-1cf6ae1f-9c3c-4101-b59a-adaa53cf4747 service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Updating instance_info_cache with network_info: [{"id": "32f5d56d-a029-4469-ad3c-77a4ca51bc02", "address": "fa:16:3e:da:8c:98", "network": {"id": "036dc114-6ea0-46cb-a958-4a6137e7e85f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-728627542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e1772c9a6c4671ab89df2e8b40cae1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32f5d56d-a0", "ovs_interfaceid": "32f5d56d-a029-4469-ad3c-77a4ca51bc02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.272451] env[62600]: DEBUG oslo_vmware.api [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222754, 'name': CloneVM_Task} progress is 94%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.296768] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52773993-9ff3-3446-6fb7-2ab4cd5a4a9f, 'name': SearchDatastore_Task, 'duration_secs': 0.008731} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.297098] env[62600]: DEBUG oslo_concurrency.lockutils [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.297344] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 927.297585] env[62600]: DEBUG oslo_concurrency.lockutils [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.297739] env[62600]: DEBUG oslo_concurrency.lockutils [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.297922] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 927.298205] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c8119649-2d72-4a8b-a46b-1cbaeba09360 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.312241] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 927.312493] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 927.313237] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-083f6df2-18f6-497e-b5a6-ccf183a2b6ae {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.318892] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Waiting for the task: (returnval){ [ 927.318892] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5215cadf-5bf4-f326-1652-1a91b0f3db96" [ 927.318892] env[62600]: _type = "Task" [ 927.318892] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.328195] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5215cadf-5bf4-f326-1652-1a91b0f3db96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.509019] env[62600]: ERROR nova.scheduler.client.report [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] [req-c5b6627e-ff08-480c-9789-59019cfa6ce7] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 664af347-7147-4bf5-9019-9ae15cb4aa82. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c5b6627e-ff08-480c-9789-59019cfa6ce7"}]} [ 927.527238] env[62600]: DEBUG nova.scheduler.client.report [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Refreshing inventories for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 927.542479] env[62600]: DEBUG nova.scheduler.client.report [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Updating ProviderTree inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 927.542739] env[62600]: DEBUG nova.compute.provider_tree [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 927.561552] env[62600]: DEBUG nova.scheduler.client.report [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Refreshing aggregate associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, aggregates: None {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 927.574702] env[62600]: DEBUG oslo_concurrency.lockutils [req-89c74122-2fb0-4065-b5d4-2c79a25d2bf0 req-e8cd5025-1a27-4fac-9af9-5d5d08a0296e service nova] Releasing lock "refresh_cache-c2f38d56-72fa-4bf0-a2cb-13d666571466" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.582588] env[62600]: DEBUG nova.compute.manager [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 927.586851] env[62600]: DEBUG nova.scheduler.client.report [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Refreshing trait associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 927.626855] env[62600]: DEBUG oslo_concurrency.lockutils [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "ab11cdd1-80a7-43c4-afac-24c7930238c1" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.627136] env[62600]: DEBUG oslo_concurrency.lockutils [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "ab11cdd1-80a7-43c4-afac-24c7930238c1" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.639997] env[62600]: DEBUG nova.compute.manager [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 927.717477] env[62600]: DEBUG oslo_concurrency.lockutils [req-fd553666-242a-4c67-a38b-3b50ad501110 req-1cf6ae1f-9c3c-4101-b59a-adaa53cf4747 service nova] Releasing lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.777145] env[62600]: DEBUG oslo_vmware.api [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222754, 'name': CloneVM_Task} progress is 94%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.832667] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5215cadf-5bf4-f326-1652-1a91b0f3db96, 'name': SearchDatastore_Task, 'duration_secs': 0.014805} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.833455] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3a467a9-ead5-42b5-a7f3-6e3b1096cf50 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.840419] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Waiting for the task: (returnval){ [ 927.840419] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b64c79-b386-617c-9f8d-fd84dc0ce762" [ 927.840419] env[62600]: _type = "Task" [ 927.840419] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.849938] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b64c79-b386-617c-9f8d-fd84dc0ce762, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.865851] env[62600]: DEBUG nova.compute.manager [req-e37d7b47-6a96-47fd-b157-15cd35d00b4b req-dda9769b-b596-4a95-a2e2-f29abd9897e2 service nova] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Received event network-vif-deleted-b4b6687a-b518-40c1-99a5-7d1ffe49e1f2 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.866082] env[62600]: INFO nova.compute.manager [req-e37d7b47-6a96-47fd-b157-15cd35d00b4b req-dda9769b-b596-4a95-a2e2-f29abd9897e2 service nova] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Neutron deleted interface b4b6687a-b518-40c1-99a5-7d1ffe49e1f2; detaching it from the instance and deleting it from the info cache [ 927.866421] env[62600]: DEBUG nova.network.neutron [req-e37d7b47-6a96-47fd-b157-15cd35d00b4b req-dda9769b-b596-4a95-a2e2-f29abd9897e2 service nova] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.916805] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57a25460-cdc4-4956-a10c-272b65ab93f5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.925869] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4448f52c-ea75-4ee1-8c10-f800ba0bc74e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.957992] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73860803-a582-4ea3-8a2b-f9088affeb0f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.965924] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-422a5c2b-c74c-48e8-aafe-066df6471162 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.980796] env[62600]: DEBUG nova.compute.provider_tree [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 928.105160] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.133574] env[62600]: DEBUG nova.compute.utils [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 928.160489] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.171759] env[62600]: DEBUG nova.network.neutron [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Successfully updated port: 3a8aca44-9791-4493-a1fd-114c9f20fb3f {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 928.255796] env[62600]: DEBUG nova.network.neutron [-] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.272616] env[62600]: DEBUG oslo_vmware.api [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222754, 'name': CloneVM_Task} progress is 94%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.353662] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b64c79-b386-617c-9f8d-fd84dc0ce762, 'name': SearchDatastore_Task, 'duration_secs': 0.009256} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.353662] env[62600]: DEBUG oslo_concurrency.lockutils [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.353662] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] c2f38d56-72fa-4bf0-a2cb-13d666571466/c2f38d56-72fa-4bf0-a2cb-13d666571466.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 928.353662] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cb9314a3-b6e1-44f7-b03f-71ac51cccb90 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.364900] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Waiting for the task: (returnval){ [ 928.364900] env[62600]: value = "task-1222755" [ 928.364900] env[62600]: _type = "Task" [ 928.364900] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.370842] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c1a7d20-f397-4d04-a292-985bfa9f7fae {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.376375] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': task-1222755, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.386163] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c3e4561-455b-455b-9f64-9c0044469db4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.416726] env[62600]: DEBUG nova.compute.manager [req-e37d7b47-6a96-47fd-b157-15cd35d00b4b req-dda9769b-b596-4a95-a2e2-f29abd9897e2 service nova] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Detach interface failed, port_id=b4b6687a-b518-40c1-99a5-7d1ffe49e1f2, reason: Instance afe838cc-e086-4986-87ec-4e1266bcaf60 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 928.530831] env[62600]: DEBUG nova.scheduler.client.report [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 110 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 928.531217] env[62600]: DEBUG nova.compute.provider_tree [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 110 to 111 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 928.531466] env[62600]: DEBUG nova.compute.provider_tree [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 928.635997] env[62600]: DEBUG oslo_concurrency.lockutils [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "ab11cdd1-80a7-43c4-afac-24c7930238c1" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.673673] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.673878] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.674112] env[62600]: DEBUG nova.network.neutron [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 928.759259] env[62600]: INFO nova.compute.manager [-] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Took 1.55 seconds to deallocate network for instance. [ 928.775516] env[62600]: DEBUG oslo_vmware.api [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222754, 'name': CloneVM_Task} progress is 95%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.876709] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': task-1222755, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463997} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.877309] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] c2f38d56-72fa-4bf0-a2cb-13d666571466/c2f38d56-72fa-4bf0-a2cb-13d666571466.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 928.877309] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 928.877473] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9a188004-92ae-440a-93bb-f015f0b44ed5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.883921] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Waiting for the task: (returnval){ [ 928.883921] env[62600]: value = "task-1222756" [ 928.883921] env[62600]: _type = "Task" [ 928.883921] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.891483] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': task-1222756, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.038884] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.979s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.041224] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.823s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.042785] env[62600]: INFO nova.compute.claims [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 929.059572] env[62600]: INFO nova.scheduler.client.report [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Deleted allocations for instance 4eaa22b9-5834-4425-82a9-d16e316f3a52 [ 929.206543] env[62600]: DEBUG nova.network.neutron [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 929.271014] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.277165] env[62600]: DEBUG oslo_vmware.api [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222754, 'name': CloneVM_Task, 'duration_secs': 2.162651} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.277415] env[62600]: INFO nova.virt.vmwareapi.vmops [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Created linked-clone VM from snapshot [ 929.278140] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58eb6727-8151-4eb9-8b68-ddbfe383ae98 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.284994] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Uploading image de4b4dc6-9859-4765-a3c0-223657c111a4 {{(pid=62600) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 929.310741] env[62600]: DEBUG oslo_vmware.rw_handles [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 929.310741] env[62600]: value = "vm-264334" [ 929.310741] env[62600]: _type = "VirtualMachine" [ 929.310741] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 929.311432] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-6db39f75-e161-4b47-957f-0204c3fd2afc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.318216] env[62600]: DEBUG oslo_vmware.rw_handles [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Lease: (returnval){ [ 929.318216] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52bc67ad-1759-71b1-2487-8003de5a8bcd" [ 929.318216] env[62600]: _type = "HttpNfcLease" [ 929.318216] env[62600]: } obtained for exporting VM: (result){ [ 929.318216] env[62600]: value = "vm-264334" [ 929.318216] env[62600]: _type = "VirtualMachine" [ 929.318216] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 929.318503] env[62600]: DEBUG oslo_vmware.api [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for the lease: (returnval){ [ 929.318503] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52bc67ad-1759-71b1-2487-8003de5a8bcd" [ 929.318503] env[62600]: _type = "HttpNfcLease" [ 929.318503] env[62600]: } to be ready. {{(pid=62600) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 929.326255] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 929.326255] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52bc67ad-1759-71b1-2487-8003de5a8bcd" [ 929.326255] env[62600]: _type = "HttpNfcLease" [ 929.326255] env[62600]: } is initializing. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 929.342164] env[62600]: DEBUG nova.network.neutron [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Updating instance_info_cache with network_info: [{"id": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "address": "fa:16:3e:67:24:01", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8aca44-97", "ovs_interfaceid": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.394299] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': task-1222756, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066439} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.394582] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 929.395426] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b0d7a76-60c7-4e03-bed2-ba56e1c96ddd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.418522] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] c2f38d56-72fa-4bf0-a2cb-13d666571466/c2f38d56-72fa-4bf0-a2cb-13d666571466.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 929.418898] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e971624c-41c8-42d3-844e-9066ec5a0c73 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.438709] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Waiting for the task: (returnval){ [ 929.438709] env[62600]: value = "task-1222758" [ 929.438709] env[62600]: _type = "Task" [ 929.438709] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.447194] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': task-1222758, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.566450] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3730cd55-8af5-4dd4-bd6c-f77f44a0eb6b tempest-ServerGroupTestJSON-2075159514 tempest-ServerGroupTestJSON-2075159514-project-member] Lock "4eaa22b9-5834-4425-82a9-d16e316f3a52" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.490s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.696771] env[62600]: DEBUG oslo_concurrency.lockutils [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "ab11cdd1-80a7-43c4-afac-24c7930238c1" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.696771] env[62600]: DEBUG oslo_concurrency.lockutils [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "ab11cdd1-80a7-43c4-afac-24c7930238c1" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.696771] env[62600]: INFO nova.compute.manager [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Attaching volume c49c9652-3f74-46a1-a176-6b393fc416ba to /dev/sdb [ 929.730488] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e3db799-df5a-4bb0-a286-2b17eb694eb4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.739807] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a0c3a2-3383-45fa-bd58-8b18395a32a3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.752745] env[62600]: DEBUG nova.virt.block_device [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Updating existing volume attachment record: 9248b461-fb2d-4961-b796-296c0b7cba0f {{(pid=62600) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 929.826256] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 929.826256] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52bc67ad-1759-71b1-2487-8003de5a8bcd" [ 929.826256] env[62600]: _type = "HttpNfcLease" [ 929.826256] env[62600]: } is ready. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 929.826641] env[62600]: DEBUG oslo_vmware.rw_handles [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 929.826641] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52bc67ad-1759-71b1-2487-8003de5a8bcd" [ 929.826641] env[62600]: _type = "HttpNfcLease" [ 929.826641] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 929.827354] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31f5c6f9-e252-48b5-8737-d5ed8b7d43ab {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.835018] env[62600]: DEBUG oslo_vmware.rw_handles [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528b9f08-ad03-613d-4b95-fe7de85275c9/disk-0.vmdk from lease info. {{(pid=62600) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 929.835217] env[62600]: DEBUG oslo_vmware.rw_handles [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528b9f08-ad03-613d-4b95-fe7de85275c9/disk-0.vmdk for reading. {{(pid=62600) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 929.894118] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.894452] env[62600]: DEBUG nova.compute.manager [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Instance network_info: |[{"id": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "address": "fa:16:3e:67:24:01", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8aca44-97", "ovs_interfaceid": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 929.896203] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:67:24:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '184687d6-125a-4b58-bb5b-fdb404088eda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3a8aca44-9791-4493-a1fd-114c9f20fb3f', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 929.903944] env[62600]: DEBUG oslo.service.loopingcall [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 929.905538] env[62600]: DEBUG nova.compute.manager [req-c0a32428-d18c-4ae9-8d0e-faef7c1c2b22 req-f3b06ace-b015-4b65-b236-1bce541c610c service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Received event network-vif-plugged-3a8aca44-9791-4493-a1fd-114c9f20fb3f {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.905797] env[62600]: DEBUG oslo_concurrency.lockutils [req-c0a32428-d18c-4ae9-8d0e-faef7c1c2b22 req-f3b06ace-b015-4b65-b236-1bce541c610c service nova] Acquiring lock "88cf521d-f4a3-493a-894d-c7f799dc0443-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.905956] env[62600]: DEBUG oslo_concurrency.lockutils [req-c0a32428-d18c-4ae9-8d0e-faef7c1c2b22 req-f3b06ace-b015-4b65-b236-1bce541c610c service nova] Lock "88cf521d-f4a3-493a-894d-c7f799dc0443-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.906150] env[62600]: DEBUG oslo_concurrency.lockutils [req-c0a32428-d18c-4ae9-8d0e-faef7c1c2b22 req-f3b06ace-b015-4b65-b236-1bce541c610c service nova] Lock "88cf521d-f4a3-493a-894d-c7f799dc0443-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.906327] env[62600]: DEBUG nova.compute.manager [req-c0a32428-d18c-4ae9-8d0e-faef7c1c2b22 req-f3b06ace-b015-4b65-b236-1bce541c610c service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] No waiting events found dispatching network-vif-plugged-3a8aca44-9791-4493-a1fd-114c9f20fb3f {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 929.906497] env[62600]: WARNING nova.compute.manager [req-c0a32428-d18c-4ae9-8d0e-faef7c1c2b22 req-f3b06ace-b015-4b65-b236-1bce541c610c service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Received unexpected event network-vif-plugged-3a8aca44-9791-4493-a1fd-114c9f20fb3f for instance with vm_state building and task_state spawning. [ 929.906663] env[62600]: DEBUG nova.compute.manager [req-c0a32428-d18c-4ae9-8d0e-faef7c1c2b22 req-f3b06ace-b015-4b65-b236-1bce541c610c service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Received event network-changed-3a8aca44-9791-4493-a1fd-114c9f20fb3f {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 929.906823] env[62600]: DEBUG nova.compute.manager [req-c0a32428-d18c-4ae9-8d0e-faef7c1c2b22 req-f3b06ace-b015-4b65-b236-1bce541c610c service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Refreshing instance network info cache due to event network-changed-3a8aca44-9791-4493-a1fd-114c9f20fb3f. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 929.907074] env[62600]: DEBUG oslo_concurrency.lockutils [req-c0a32428-d18c-4ae9-8d0e-faef7c1c2b22 req-f3b06ace-b015-4b65-b236-1bce541c610c service nova] Acquiring lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.907301] env[62600]: DEBUG oslo_concurrency.lockutils [req-c0a32428-d18c-4ae9-8d0e-faef7c1c2b22 req-f3b06ace-b015-4b65-b236-1bce541c610c service nova] Acquired lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.907455] env[62600]: DEBUG nova.network.neutron [req-c0a32428-d18c-4ae9-8d0e-faef7c1c2b22 req-f3b06ace-b015-4b65-b236-1bce541c610c service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Refreshing network info cache for port 3a8aca44-9791-4493-a1fd-114c9f20fb3f {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 929.910170] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 929.910671] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0cdeade3-0a32-4a37-8129-d2af80e06d98 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.933769] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 929.933769] env[62600]: value = "task-1222760" [ 929.933769] env[62600]: _type = "Task" [ 929.933769] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.942894] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222760, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.951681] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': task-1222758, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.030322] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-68a77e5c-7a2d-4f03-bc93-c87c71f22370 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.317888] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbca0f62-54f7-4f0b-a03c-83d8c4d7330f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.325491] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e8e595-0534-4b36-961a-3d8028fd6d85 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.360190] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-313fcd22-6862-45b6-9a77-8920740368e4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.368858] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-727d9134-0c84-4fab-ab5b-ca0bfa16ef4a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.384149] env[62600]: DEBUG nova.compute.provider_tree [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.446821] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222760, 'name': CreateVM_Task, 'duration_secs': 0.383731} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.452468] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 930.453315] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.453549] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.454165] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 930.455032] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6a7572c-feb6-42e9-ae25-275d5afb36a5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.460917] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': task-1222758, 'name': ReconfigVM_Task, 'duration_secs': 0.802009} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.463974] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Reconfigured VM instance instance-00000051 to attach disk [datastore2] c2f38d56-72fa-4bf0-a2cb-13d666571466/c2f38d56-72fa-4bf0-a2cb-13d666571466.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 930.463974] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fcaa2689-2065-42b5-839c-59c811c7173d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.465551] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 930.465551] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52df886c-9b40-fc48-22b6-ea00e7f9f53c" [ 930.465551] env[62600]: _type = "Task" [ 930.465551] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.470683] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Waiting for the task: (returnval){ [ 930.470683] env[62600]: value = "task-1222763" [ 930.470683] env[62600]: _type = "Task" [ 930.470683] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.479940] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52df886c-9b40-fc48-22b6-ea00e7f9f53c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.484247] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': task-1222763, 'name': Rename_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.669402] env[62600]: DEBUG nova.network.neutron [req-c0a32428-d18c-4ae9-8d0e-faef7c1c2b22 req-f3b06ace-b015-4b65-b236-1bce541c610c service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Updated VIF entry in instance network info cache for port 3a8aca44-9791-4493-a1fd-114c9f20fb3f. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 930.669978] env[62600]: DEBUG nova.network.neutron [req-c0a32428-d18c-4ae9-8d0e-faef7c1c2b22 req-f3b06ace-b015-4b65-b236-1bce541c610c service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Updating instance_info_cache with network_info: [{"id": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "address": "fa:16:3e:67:24:01", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8aca44-97", "ovs_interfaceid": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.887216] env[62600]: DEBUG nova.scheduler.client.report [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.979340] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52df886c-9b40-fc48-22b6-ea00e7f9f53c, 'name': SearchDatastore_Task, 'duration_secs': 0.011899} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.980240] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.981073] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 930.981073] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.981073] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.983200] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 930.984688] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-792e1786-5436-46ea-8a15-bb910a5bf70f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.986486] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': task-1222763, 'name': Rename_Task, 'duration_secs': 0.220394} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.986742] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 930.987384] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-786d36b1-df79-427c-9f5a-00c7b608e87f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.994623] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 930.994623] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 930.994964] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef3bb402-f737-452a-bc4c-cd5cd912988c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.997390] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Waiting for the task: (returnval){ [ 930.997390] env[62600]: value = "task-1222764" [ 930.997390] env[62600]: _type = "Task" [ 930.997390] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.002023] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 931.002023] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d78c7d-7cee-a513-3b0b-146cc5a63ccd" [ 931.002023] env[62600]: _type = "Task" [ 931.002023] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.009587] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': task-1222764, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.014968] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d78c7d-7cee-a513-3b0b-146cc5a63ccd, 'name': SearchDatastore_Task, 'duration_secs': 0.009202} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.015746] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3282db0-b467-4e43-bc64-e62e5c0f5163 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.021630] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 931.021630] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521e4f99-c47b-4ccc-6a62-15b24fb7f26d" [ 931.021630] env[62600]: _type = "Task" [ 931.021630] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.030655] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521e4f99-c47b-4ccc-6a62-15b24fb7f26d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.177934] env[62600]: DEBUG oslo_concurrency.lockutils [req-c0a32428-d18c-4ae9-8d0e-faef7c1c2b22 req-f3b06ace-b015-4b65-b236-1bce541c610c service nova] Releasing lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.395593] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.354s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.396209] env[62600]: DEBUG nova.compute.manager [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 931.399104] env[62600]: DEBUG oslo_concurrency.lockutils [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.860s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.399467] env[62600]: DEBUG nova.objects.instance [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lazy-loading 'resources' on Instance uuid d627a701-77cd-4a1d-9e52-0fc4adc93391 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.508334] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': task-1222764, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.532865] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521e4f99-c47b-4ccc-6a62-15b24fb7f26d, 'name': SearchDatastore_Task, 'duration_secs': 0.010111} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.533185] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.533587] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 88cf521d-f4a3-493a-894d-c7f799dc0443/88cf521d-f4a3-493a-894d-c7f799dc0443.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 931.533886] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7cd764cc-069a-41bc-83b9-d0f59a4eb812 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.543376] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 931.543376] env[62600]: value = "task-1222765" [ 931.543376] env[62600]: _type = "Task" [ 931.543376] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.553215] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222765, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.903746] env[62600]: DEBUG nova.compute.utils [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 931.905646] env[62600]: DEBUG nova.compute.manager [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 931.905880] env[62600]: DEBUG nova.network.neutron [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 931.976332] env[62600]: DEBUG nova.policy [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bd0d91fa2ac74c38b9e0ebf39c3baa7a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73e664d178f7484a9f4741b4d9450e68', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 932.018288] env[62600]: DEBUG oslo_vmware.api [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': task-1222764, 'name': PowerOnVM_Task, 'duration_secs': 0.656663} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.018921] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 932.019372] env[62600]: INFO nova.compute.manager [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Took 8.31 seconds to spawn the instance on the hypervisor. [ 932.019827] env[62600]: DEBUG nova.compute.manager [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.021309] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ee72fa-ab52-4dbf-b70f-ebd7fc1faac6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.062773] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222765, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.237344] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b931c8b0-24f6-45e1-bbb0-b1cad8c986f5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.246245] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ec18d6-2342-43fc-b0b8-b99cccb6ce8c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.287284] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45482a3d-feda-4c47-8b66-4b4a379f6bd5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.297088] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-704353e8-7472-4417-9547-0006405e194a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.313628] env[62600]: DEBUG nova.compute.provider_tree [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 932.335101] env[62600]: DEBUG nova.network.neutron [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Successfully created port: 874235d5-d1b2-4382-a30a-2e187141de5b {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 932.411655] env[62600]: DEBUG nova.compute.manager [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 932.554626] env[62600]: INFO nova.compute.manager [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Took 27.49 seconds to build instance. [ 932.560020] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222765, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.598921} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.560306] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 88cf521d-f4a3-493a-894d-c7f799dc0443/88cf521d-f4a3-493a-894d-c7f799dc0443.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 932.560529] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 932.560794] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0c4fdb69-19b8-4aac-9541-c0a5dc8b728b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.567916] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 932.567916] env[62600]: value = "task-1222767" [ 932.567916] env[62600]: _type = "Task" [ 932.567916] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.582768] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222767, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.847968] env[62600]: DEBUG nova.scheduler.client.report [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 111 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 932.848279] env[62600]: DEBUG nova.compute.provider_tree [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 111 to 112 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 932.848485] env[62600]: DEBUG nova.compute.provider_tree [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 933.056697] env[62600]: DEBUG oslo_concurrency.lockutils [None req-11b1217a-b8ea-4ee2-9ede-0efaf0e4c039 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Lock "c2f38d56-72fa-4bf0-a2cb-13d666571466" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.077710] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222767, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.133433} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.077990] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 933.078821] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f99d6656-f77f-432c-abb4-8ae6363815ba {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.101104] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 88cf521d-f4a3-493a-894d-c7f799dc0443/88cf521d-f4a3-493a-894d-c7f799dc0443.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 933.101460] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23bb24ea-ad20-4206-a8aa-4d0571158061 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.121023] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 933.121023] env[62600]: value = "task-1222768" [ 933.121023] env[62600]: _type = "Task" [ 933.121023] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.130168] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222768, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.336830] env[62600]: DEBUG nova.compute.manager [req-fa522683-8282-45c4-91f3-b83f9a0a8e73 req-4efad86c-090e-41c6-a715-113088219abf service nova] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Received event network-changed-c715a88c-6498-428e-989d-3a1aa2f2e3bd {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.337170] env[62600]: DEBUG nova.compute.manager [req-fa522683-8282-45c4-91f3-b83f9a0a8e73 req-4efad86c-090e-41c6-a715-113088219abf service nova] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Refreshing instance network info cache due to event network-changed-c715a88c-6498-428e-989d-3a1aa2f2e3bd. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 933.337516] env[62600]: DEBUG oslo_concurrency.lockutils [req-fa522683-8282-45c4-91f3-b83f9a0a8e73 req-4efad86c-090e-41c6-a715-113088219abf service nova] Acquiring lock "refresh_cache-c2f38d56-72fa-4bf0-a2cb-13d666571466" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.337717] env[62600]: DEBUG oslo_concurrency.lockutils [req-fa522683-8282-45c4-91f3-b83f9a0a8e73 req-4efad86c-090e-41c6-a715-113088219abf service nova] Acquired lock "refresh_cache-c2f38d56-72fa-4bf0-a2cb-13d666571466" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.337901] env[62600]: DEBUG nova.network.neutron [req-fa522683-8282-45c4-91f3-b83f9a0a8e73 req-4efad86c-090e-41c6-a715-113088219abf service nova] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Refreshing network info cache for port c715a88c-6498-428e-989d-3a1aa2f2e3bd {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 933.353672] env[62600]: DEBUG oslo_concurrency.lockutils [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.955s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.357205] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.567s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.357205] env[62600]: DEBUG nova.objects.instance [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lazy-loading 'resources' on Instance uuid 043ac602-6c4a-4275-b019-5d717579e7c0 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.379868] env[62600]: INFO nova.scheduler.client.report [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleted allocations for instance d627a701-77cd-4a1d-9e52-0fc4adc93391 [ 933.423956] env[62600]: DEBUG nova.compute.manager [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 933.448943] env[62600]: DEBUG nova.virt.hardware [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 933.449517] env[62600]: DEBUG nova.virt.hardware [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 933.452045] env[62600]: DEBUG nova.virt.hardware [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 933.452398] env[62600]: DEBUG nova.virt.hardware [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 933.452687] env[62600]: DEBUG nova.virt.hardware [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 933.455289] env[62600]: DEBUG nova.virt.hardware [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 933.455289] env[62600]: DEBUG nova.virt.hardware [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 933.455289] env[62600]: DEBUG nova.virt.hardware [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 933.455289] env[62600]: DEBUG nova.virt.hardware [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 933.455289] env[62600]: DEBUG nova.virt.hardware [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 933.455289] env[62600]: DEBUG nova.virt.hardware [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 933.456524] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2abe5c3c-6fff-4535-8705-86b045b718e0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.467045] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1509ea21-9c43-4539-8881-9036a42e15f2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.631076] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222768, 'name': ReconfigVM_Task, 'duration_secs': 0.506721} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.631459] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 88cf521d-f4a3-493a-894d-c7f799dc0443/88cf521d-f4a3-493a-894d-c7f799dc0443.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 933.632132] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3e8d602c-14b1-4c81-8243-5b6ab536df41 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.641036] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 933.641036] env[62600]: value = "task-1222769" [ 933.641036] env[62600]: _type = "Task" [ 933.641036] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.650137] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222769, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.891655] env[62600]: DEBUG oslo_concurrency.lockutils [None req-42614b01-8b6e-427f-ad8f-415c84aa2403 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "d627a701-77cd-4a1d-9e52-0fc4adc93391" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.285s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.931876] env[62600]: DEBUG nova.network.neutron [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Successfully updated port: 874235d5-d1b2-4382-a30a-2e187141de5b {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 934.105544] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7efbd404-1d8c-4f70-98c0-14303ee855f8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.113989] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d951066-3b0c-4ba4-968e-aa8af9a99f78 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.118938] env[62600]: DEBUG nova.network.neutron [req-fa522683-8282-45c4-91f3-b83f9a0a8e73 req-4efad86c-090e-41c6-a715-113088219abf service nova] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Updated VIF entry in instance network info cache for port c715a88c-6498-428e-989d-3a1aa2f2e3bd. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 934.119341] env[62600]: DEBUG nova.network.neutron [req-fa522683-8282-45c4-91f3-b83f9a0a8e73 req-4efad86c-090e-41c6-a715-113088219abf service nova] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Updating instance_info_cache with network_info: [{"id": "c715a88c-6498-428e-989d-3a1aa2f2e3bd", "address": "fa:16:3e:67:c1:e4", "network": {"id": "7ed8725a-d4a1-4d66-8bd3-08c9abf86064", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1314620504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4e088a25c01c40d98d27e22563b4c625", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b98c49ac-0eb7-4311-aa8f-60581b2ce706", "external-id": "nsx-vlan-transportzone-184", "segmentation_id": 184, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc715a88c-64", "ovs_interfaceid": "c715a88c-6498-428e-989d-3a1aa2f2e3bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.146879] env[62600]: DEBUG oslo_concurrency.lockutils [req-fa522683-8282-45c4-91f3-b83f9a0a8e73 req-4efad86c-090e-41c6-a715-113088219abf service nova] Releasing lock "refresh_cache-c2f38d56-72fa-4bf0-a2cb-13d666571466" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.150607] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-893a29b3-b3d1-4589-b20d-456e4be1c8f4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.160908] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222769, 'name': Rename_Task, 'duration_secs': 0.156909} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.162069] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b2c808-5960-45cd-b032-2191660d1291 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.165670] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 934.166151] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4edfb247-5831-451a-a43f-63b6ee85335c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.179154] env[62600]: DEBUG nova.compute.provider_tree [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.181537] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 934.181537] env[62600]: value = "task-1222770" [ 934.181537] env[62600]: _type = "Task" [ 934.181537] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.190720] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222770, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.299905] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Volume attach. Driver type: vmdk {{(pid=62600) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 934.300211] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264337', 'volume_id': 'c49c9652-3f74-46a1-a176-6b393fc416ba', 'name': 'volume-c49c9652-3f74-46a1-a176-6b393fc416ba', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ab11cdd1-80a7-43c4-afac-24c7930238c1', 'attached_at': '', 'detached_at': '', 'volume_id': 'c49c9652-3f74-46a1-a176-6b393fc416ba', 'serial': 'c49c9652-3f74-46a1-a176-6b393fc416ba'} {{(pid=62600) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 934.301150] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4eb8fb5-bdd9-44d3-8a02-b0280284a6d4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.318857] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf4678c-c8df-4f8d-acb9-f6d7ef836f29 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.345843] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] volume-c49c9652-3f74-46a1-a176-6b393fc416ba/volume-c49c9652-3f74-46a1-a176-6b393fc416ba.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 934.346236] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6eac3cf6-b468-42f0-ae1c-08d07a979df6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.367605] env[62600]: DEBUG oslo_vmware.api [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 934.367605] env[62600]: value = "task-1222771" [ 934.367605] env[62600]: _type = "Task" [ 934.367605] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.377181] env[62600]: DEBUG oslo_vmware.api [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222771, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.435310] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "refresh_cache-f18df958-2885-4d8c-a9ad-140faaed1178" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.435575] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "refresh_cache-f18df958-2885-4d8c-a9ad-140faaed1178" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.435878] env[62600]: DEBUG nova.network.neutron [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 934.683779] env[62600]: DEBUG nova.scheduler.client.report [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 934.699988] env[62600]: DEBUG oslo_vmware.api [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222770, 'name': PowerOnVM_Task, 'duration_secs': 0.462576} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.701199] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 934.701472] env[62600]: INFO nova.compute.manager [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Took 7.63 seconds to spawn the instance on the hypervisor. [ 934.701665] env[62600]: DEBUG nova.compute.manager [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 934.702610] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0950388-bf1b-43dd-814c-913ff00d0efe {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.877848] env[62600]: DEBUG oslo_vmware.api [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222771, 'name': ReconfigVM_Task, 'duration_secs': 0.431736} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.878168] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Reconfigured VM instance instance-00000046 to attach disk [datastore2] volume-c49c9652-3f74-46a1-a176-6b393fc416ba/volume-c49c9652-3f74-46a1-a176-6b393fc416ba.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 934.882844] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a9c0a80-e8c6-4a24-b590-e4c014b5ffe6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.899740] env[62600]: DEBUG oslo_vmware.api [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 934.899740] env[62600]: value = "task-1222772" [ 934.899740] env[62600]: _type = "Task" [ 934.899740] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.910364] env[62600]: DEBUG oslo_vmware.api [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222772, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.970675] env[62600]: DEBUG nova.network.neutron [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 935.117457] env[62600]: DEBUG nova.network.neutron [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Updating instance_info_cache with network_info: [{"id": "874235d5-d1b2-4382-a30a-2e187141de5b", "address": "fa:16:3e:73:fa:2c", "network": {"id": "95818650-8227-4d11-9d14-007f06c07790", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1738152115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73e664d178f7484a9f4741b4d9450e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap874235d5-d1", "ovs_interfaceid": "874235d5-d1b2-4382-a30a-2e187141de5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.194231] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.838s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.196620] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.331s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.198106] env[62600]: INFO nova.compute.claims [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 935.218724] env[62600]: INFO nova.scheduler.client.report [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Deleted allocations for instance 043ac602-6c4a-4275-b019-5d717579e7c0 [ 935.224266] env[62600]: INFO nova.compute.manager [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Took 25.04 seconds to build instance. [ 935.362731] env[62600]: DEBUG nova.compute.manager [req-9a834df1-8283-43ce-9190-3c77f357cc8c req-18dd035b-1d8f-4e43-9945-647b835bdd05 service nova] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Received event network-vif-plugged-874235d5-d1b2-4382-a30a-2e187141de5b {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.362923] env[62600]: DEBUG oslo_concurrency.lockutils [req-9a834df1-8283-43ce-9190-3c77f357cc8c req-18dd035b-1d8f-4e43-9945-647b835bdd05 service nova] Acquiring lock "f18df958-2885-4d8c-a9ad-140faaed1178-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.363213] env[62600]: DEBUG oslo_concurrency.lockutils [req-9a834df1-8283-43ce-9190-3c77f357cc8c req-18dd035b-1d8f-4e43-9945-647b835bdd05 service nova] Lock "f18df958-2885-4d8c-a9ad-140faaed1178-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.363452] env[62600]: DEBUG oslo_concurrency.lockutils [req-9a834df1-8283-43ce-9190-3c77f357cc8c req-18dd035b-1d8f-4e43-9945-647b835bdd05 service nova] Lock "f18df958-2885-4d8c-a9ad-140faaed1178-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.363576] env[62600]: DEBUG nova.compute.manager [req-9a834df1-8283-43ce-9190-3c77f357cc8c req-18dd035b-1d8f-4e43-9945-647b835bdd05 service nova] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] No waiting events found dispatching network-vif-plugged-874235d5-d1b2-4382-a30a-2e187141de5b {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 935.363747] env[62600]: WARNING nova.compute.manager [req-9a834df1-8283-43ce-9190-3c77f357cc8c req-18dd035b-1d8f-4e43-9945-647b835bdd05 service nova] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Received unexpected event network-vif-plugged-874235d5-d1b2-4382-a30a-2e187141de5b for instance with vm_state building and task_state spawning. [ 935.364016] env[62600]: DEBUG nova.compute.manager [req-9a834df1-8283-43ce-9190-3c77f357cc8c req-18dd035b-1d8f-4e43-9945-647b835bdd05 service nova] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Received event network-changed-874235d5-d1b2-4382-a30a-2e187141de5b {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.364213] env[62600]: DEBUG nova.compute.manager [req-9a834df1-8283-43ce-9190-3c77f357cc8c req-18dd035b-1d8f-4e43-9945-647b835bdd05 service nova] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Refreshing instance network info cache due to event network-changed-874235d5-d1b2-4382-a30a-2e187141de5b. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 935.364391] env[62600]: DEBUG oslo_concurrency.lockutils [req-9a834df1-8283-43ce-9190-3c77f357cc8c req-18dd035b-1d8f-4e43-9945-647b835bdd05 service nova] Acquiring lock "refresh_cache-f18df958-2885-4d8c-a9ad-140faaed1178" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.409947] env[62600]: DEBUG oslo_vmware.api [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222772, 'name': ReconfigVM_Task, 'duration_secs': 0.152459} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.410235] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264337', 'volume_id': 'c49c9652-3f74-46a1-a176-6b393fc416ba', 'name': 'volume-c49c9652-3f74-46a1-a176-6b393fc416ba', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ab11cdd1-80a7-43c4-afac-24c7930238c1', 'attached_at': '', 'detached_at': '', 'volume_id': 'c49c9652-3f74-46a1-a176-6b393fc416ba', 'serial': 'c49c9652-3f74-46a1-a176-6b393fc416ba'} {{(pid=62600) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 935.620048] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "refresh_cache-f18df958-2885-4d8c-a9ad-140faaed1178" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.620488] env[62600]: DEBUG nova.compute.manager [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Instance network_info: |[{"id": "874235d5-d1b2-4382-a30a-2e187141de5b", "address": "fa:16:3e:73:fa:2c", "network": {"id": "95818650-8227-4d11-9d14-007f06c07790", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1738152115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73e664d178f7484a9f4741b4d9450e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap874235d5-d1", "ovs_interfaceid": "874235d5-d1b2-4382-a30a-2e187141de5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 935.620956] env[62600]: DEBUG oslo_concurrency.lockutils [req-9a834df1-8283-43ce-9190-3c77f357cc8c req-18dd035b-1d8f-4e43-9945-647b835bdd05 service nova] Acquired lock "refresh_cache-f18df958-2885-4d8c-a9ad-140faaed1178" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.621170] env[62600]: DEBUG nova.network.neutron [req-9a834df1-8283-43ce-9190-3c77f357cc8c req-18dd035b-1d8f-4e43-9945-647b835bdd05 service nova] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Refreshing network info cache for port 874235d5-d1b2-4382-a30a-2e187141de5b {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 935.622513] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:fa:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43798f54-0c07-4417-a23f-58bb6b7e204b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '874235d5-d1b2-4382-a30a-2e187141de5b', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 935.629993] env[62600]: DEBUG oslo.service.loopingcall [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.630750] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 935.631516] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6400a31b-d0cd-4b5e-b5af-02878ddcf368 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.652878] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 935.652878] env[62600]: value = "task-1222773" [ 935.652878] env[62600]: _type = "Task" [ 935.652878] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.662069] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222773, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.728561] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ae5afd10-8b21-4a51-bec9-1e6a00fda98b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "88cf521d-f4a3-493a-894d-c7f799dc0443" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.559s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.729138] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fa3711f5-8194-4d3d-b9a7-b8581857ca12 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "043ac602-6c4a-4275-b019-5d717579e7c0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.371s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.164654] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222773, 'name': CreateVM_Task, 'duration_secs': 0.459338} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.164872] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 936.165727] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.165887] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.166234] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 936.166504] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b911d941-4f2c-4ac9-a269-6455724af125 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.173365] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 936.173365] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52819acf-9118-e77e-3c76-ec44521a6ebd" [ 936.173365] env[62600]: _type = "Task" [ 936.173365] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.184407] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52819acf-9118-e77e-3c76-ec44521a6ebd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.365448] env[62600]: DEBUG nova.network.neutron [req-9a834df1-8283-43ce-9190-3c77f357cc8c req-18dd035b-1d8f-4e43-9945-647b835bdd05 service nova] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Updated VIF entry in instance network info cache for port 874235d5-d1b2-4382-a30a-2e187141de5b. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 936.365940] env[62600]: DEBUG nova.network.neutron [req-9a834df1-8283-43ce-9190-3c77f357cc8c req-18dd035b-1d8f-4e43-9945-647b835bdd05 service nova] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Updating instance_info_cache with network_info: [{"id": "874235d5-d1b2-4382-a30a-2e187141de5b", "address": "fa:16:3e:73:fa:2c", "network": {"id": "95818650-8227-4d11-9d14-007f06c07790", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1738152115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73e664d178f7484a9f4741b4d9450e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap874235d5-d1", "ovs_interfaceid": "874235d5-d1b2-4382-a30a-2e187141de5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.426643] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77755a4c-8e9f-40d1-8ae8-b9d5d44a816a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.435410] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf06fc94-aafe-4697-95da-dd2878ed699b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.467945] env[62600]: DEBUG nova.objects.instance [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lazy-loading 'flavor' on Instance uuid ab11cdd1-80a7-43c4-afac-24c7930238c1 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 936.470531] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3672278-ac40-43c6-be6a-0e0708c91166 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.479984] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e7e632-e384-4125-a7de-f74b5479d0d8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.496497] env[62600]: DEBUG nova.compute.provider_tree [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 936.614915] env[62600]: DEBUG nova.compute.manager [req-27801d54-0457-4da4-a049-4b125033133a req-edf107e9-0ccf-43aa-b3a8-7e444c48c571 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Received event network-changed-3a8aca44-9791-4493-a1fd-114c9f20fb3f {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.615158] env[62600]: DEBUG nova.compute.manager [req-27801d54-0457-4da4-a049-4b125033133a req-edf107e9-0ccf-43aa-b3a8-7e444c48c571 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Refreshing instance network info cache due to event network-changed-3a8aca44-9791-4493-a1fd-114c9f20fb3f. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 936.615385] env[62600]: DEBUG oslo_concurrency.lockutils [req-27801d54-0457-4da4-a049-4b125033133a req-edf107e9-0ccf-43aa-b3a8-7e444c48c571 service nova] Acquiring lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.615533] env[62600]: DEBUG oslo_concurrency.lockutils [req-27801d54-0457-4da4-a049-4b125033133a req-edf107e9-0ccf-43aa-b3a8-7e444c48c571 service nova] Acquired lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.615699] env[62600]: DEBUG nova.network.neutron [req-27801d54-0457-4da4-a049-4b125033133a req-edf107e9-0ccf-43aa-b3a8-7e444c48c571 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Refreshing network info cache for port 3a8aca44-9791-4493-a1fd-114c9f20fb3f {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 936.686597] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52819acf-9118-e77e-3c76-ec44521a6ebd, 'name': SearchDatastore_Task, 'duration_secs': 0.013599} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.686920] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.687192] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 936.687488] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.687778] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.687778] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 936.688053] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-63c4e35f-6ece-4e0e-9051-7fc8ec956682 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.701156] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 936.701270] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 936.702392] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8392e1e0-ee1d-41bf-87fc-aea265e7e22d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.711063] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 936.711063] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]523ae4bb-7628-b91c-eacf-c25567d4dc79" [ 936.711063] env[62600]: _type = "Task" [ 936.711063] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.722143] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]523ae4bb-7628-b91c-eacf-c25567d4dc79, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.870972] env[62600]: DEBUG oslo_concurrency.lockutils [req-9a834df1-8283-43ce-9190-3c77f357cc8c req-18dd035b-1d8f-4e43-9945-647b835bdd05 service nova] Releasing lock "refresh_cache-f18df958-2885-4d8c-a9ad-140faaed1178" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.979998] env[62600]: DEBUG oslo_concurrency.lockutils [None req-345b5424-8b46-46f8-a0ed-38f4e6a2e952 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "ab11cdd1-80a7-43c4-afac-24c7930238c1" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.283s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.018126] env[62600]: ERROR nova.scheduler.client.report [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [req-4ba2c631-1440-45a2-920d-a2e9582d0b68] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 664af347-7147-4bf5-9019-9ae15cb4aa82. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-4ba2c631-1440-45a2-920d-a2e9582d0b68"}]} [ 937.036092] env[62600]: DEBUG nova.scheduler.client.report [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Refreshing inventories for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 937.051746] env[62600]: DEBUG nova.scheduler.client.report [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Updating ProviderTree inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 937.051994] env[62600]: DEBUG nova.compute.provider_tree [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 937.066164] env[62600]: DEBUG nova.scheduler.client.report [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Refreshing aggregate associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, aggregates: None {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 937.091813] env[62600]: DEBUG nova.scheduler.client.report [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Refreshing trait associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 937.182375] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "234dc7f8-e2ef-4f20-b25b-dcd61ce20c57" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.182737] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "234dc7f8-e2ef-4f20-b25b-dcd61ce20c57" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.225910] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]523ae4bb-7628-b91c-eacf-c25567d4dc79, 'name': SearchDatastore_Task, 'duration_secs': 0.011774} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.230290] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5634455a-5054-4b37-b60d-316a2473a96f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.237060] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 937.237060] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52fa9996-088c-b023-d126-4389a7dcf63f" [ 937.237060] env[62600]: _type = "Task" [ 937.237060] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.251802] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52fa9996-088c-b023-d126-4389a7dcf63f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.385866] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb78870-606d-41a3-a7b6-994311ad404f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.394975] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5bbf3bc-4be4-4402-a9c8-94cc37fdb758 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.430726] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec79d6d6-4e05-44ce-8c56-d20ae6338511 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.440186] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69c14bd9-8ce8-48f6-a16f-6e1cc8225013 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.458292] env[62600]: DEBUG nova.compute.provider_tree [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 937.685788] env[62600]: DEBUG nova.compute.manager [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 937.749246] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52fa9996-088c-b023-d126-4389a7dcf63f, 'name': SearchDatastore_Task, 'duration_secs': 0.014273} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.750208] env[62600]: DEBUG nova.network.neutron [req-27801d54-0457-4da4-a049-4b125033133a req-edf107e9-0ccf-43aa-b3a8-7e444c48c571 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Updated VIF entry in instance network info cache for port 3a8aca44-9791-4493-a1fd-114c9f20fb3f. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 937.750540] env[62600]: DEBUG nova.network.neutron [req-27801d54-0457-4da4-a049-4b125033133a req-edf107e9-0ccf-43aa-b3a8-7e444c48c571 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Updating instance_info_cache with network_info: [{"id": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "address": "fa:16:3e:67:24:01", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8aca44-97", "ovs_interfaceid": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.751726] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.751991] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] f18df958-2885-4d8c-a9ad-140faaed1178/f18df958-2885-4d8c-a9ad-140faaed1178.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 937.752472] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-27302a82-b6c6-4b40-8350-83600ceacddd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.761765] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 937.761765] env[62600]: value = "task-1222774" [ 937.761765] env[62600]: _type = "Task" [ 937.761765] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.772140] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222774, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.011054] env[62600]: DEBUG nova.scheduler.client.report [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 113 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 938.011054] env[62600]: DEBUG nova.compute.provider_tree [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 113 to 114 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 938.011054] env[62600]: DEBUG nova.compute.provider_tree [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 938.212532] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.254201] env[62600]: DEBUG oslo_concurrency.lockutils [req-27801d54-0457-4da4-a049-4b125033133a req-edf107e9-0ccf-43aa-b3a8-7e444c48c571 service nova] Releasing lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.273012] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222774, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.515427] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.318s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.515733] env[62600]: DEBUG nova.compute.manager [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 938.519954] env[62600]: DEBUG oslo_concurrency.lockutils [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.337s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.519954] env[62600]: DEBUG nova.objects.instance [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lazy-loading 'resources' on Instance uuid 9d899d96-9f4f-41d1-a368-3fde5efc110a {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.772986] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222774, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.882691} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.773378] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] f18df958-2885-4d8c-a9ad-140faaed1178/f18df958-2885-4d8c-a9ad-140faaed1178.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 938.773534] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 938.773756] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-04143109-c60e-439f-9092-45b9ed4410f0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.780842] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 938.780842] env[62600]: value = "task-1222775" [ 938.780842] env[62600]: _type = "Task" [ 938.780842] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.789227] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222775, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.987432] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "39deb498-6bf0-4f3b-932b-8068fc48271e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.987432] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "39deb498-6bf0-4f3b-932b-8068fc48271e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.025441] env[62600]: DEBUG nova.compute.utils [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 939.026968] env[62600]: DEBUG nova.compute.manager [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 939.027172] env[62600]: DEBUG nova.network.neutron [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 939.077584] env[62600]: DEBUG nova.policy [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '46e18992b34d412e8e3e8205d1d18265', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '734eda84fbe4493ba1f3db243a2bdb09', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 939.241405] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da12dfa-9c85-4f89-84dc-5b4951924216 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.250685] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd2090c-cb2c-4ecf-9db9-b10e6d8a4c8e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.293719] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe768cf3-c744-4582-ba2d-4baba49af454 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.301990] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222775, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062899} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.304318] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 939.305293] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87a61577-7c05-47b3-8061-b0bb184f31ad {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.308735] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e81d957-466a-476f-85cc-0849c817dc9d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.340155] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] f18df958-2885-4d8c-a9ad-140faaed1178/f18df958-2885-4d8c-a9ad-140faaed1178.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 939.340858] env[62600]: DEBUG nova.compute.provider_tree [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 939.342133] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60aa8970-dc24-4a66-83fb-8c0c52cc265c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.368946] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 939.368946] env[62600]: value = "task-1222776" [ 939.368946] env[62600]: _type = "Task" [ 939.368946] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.378676] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222776, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.390570] env[62600]: DEBUG nova.network.neutron [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Successfully created port: 64a8ba35-fe05-4c77-ac73-f420bd31a46d {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 939.489155] env[62600]: DEBUG nova.compute.manager [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 939.532263] env[62600]: DEBUG nova.compute.manager [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 939.879514] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222776, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.882759] env[62600]: ERROR nova.scheduler.client.report [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [req-ea84e5a2-9eb6-471c-ab99-ef15db0c9cff] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 664af347-7147-4bf5-9019-9ae15cb4aa82. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ea84e5a2-9eb6-471c-ab99-ef15db0c9cff"}]} [ 939.906140] env[62600]: DEBUG nova.scheduler.client.report [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Refreshing inventories for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 939.921028] env[62600]: DEBUG nova.scheduler.client.report [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Updating ProviderTree inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 939.921189] env[62600]: DEBUG nova.compute.provider_tree [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 939.934972] env[62600]: DEBUG nova.scheduler.client.report [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Refreshing aggregate associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, aggregates: None {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 939.955341] env[62600]: DEBUG nova.scheduler.client.report [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Refreshing trait associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 940.006998] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.009344] env[62600]: DEBUG oslo_vmware.rw_handles [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528b9f08-ad03-613d-4b95-fe7de85275c9/disk-0.vmdk. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 940.010042] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2920e4-5a9a-4b65-a47b-a58f25708dc0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.019593] env[62600]: DEBUG oslo_vmware.rw_handles [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528b9f08-ad03-613d-4b95-fe7de85275c9/disk-0.vmdk is in state: ready. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 940.019593] env[62600]: ERROR oslo_vmware.rw_handles [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528b9f08-ad03-613d-4b95-fe7de85275c9/disk-0.vmdk due to incomplete transfer. [ 940.019593] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-a6aa935d-b5ab-4204-bdd4-5bf86433b2fe {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.028024] env[62600]: DEBUG oslo_vmware.rw_handles [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/528b9f08-ad03-613d-4b95-fe7de85275c9/disk-0.vmdk. {{(pid=62600) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 940.028024] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Uploaded image de4b4dc6-9859-4765-a3c0-223657c111a4 to the Glance image server {{(pid=62600) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 940.031242] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Destroying the VM {{(pid=62600) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 940.033798] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-2aa0ea1e-08ac-4302-8a0f-90d3ea501fd2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.045154] env[62600]: DEBUG oslo_vmware.api [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for the task: (returnval){ [ 940.045154] env[62600]: value = "task-1222777" [ 940.045154] env[62600]: _type = "Task" [ 940.045154] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.058039] env[62600]: DEBUG oslo_vmware.api [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222777, 'name': Destroy_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.196703] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236e31b9-573d-4114-b1bb-c29b529f78d0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.204693] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dee9f82-7bb3-4cd9-a319-37d646ed0e78 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.235660] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81aa5ef1-5655-4206-99f4-97e290f78f4e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.243696] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c166e7c-bc19-49bd-a3bc-c29b8a2fc384 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.257412] env[62600]: DEBUG nova.compute.provider_tree [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 940.381773] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222776, 'name': ReconfigVM_Task, 'duration_secs': 0.541936} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.382087] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Reconfigured VM instance instance-00000053 to attach disk [datastore2] f18df958-2885-4d8c-a9ad-140faaed1178/f18df958-2885-4d8c-a9ad-140faaed1178.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 940.382739] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b6bbb8cd-e131-4aa5-8a33-7e110cb31050 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.389711] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 940.389711] env[62600]: value = "task-1222778" [ 940.389711] env[62600]: _type = "Task" [ 940.389711] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.398558] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222778, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.542057] env[62600]: DEBUG nova.compute.manager [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 940.555072] env[62600]: DEBUG oslo_vmware.api [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222777, 'name': Destroy_Task, 'duration_secs': 0.317747} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.556963] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Destroyed the VM [ 940.557241] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Deleting Snapshot of the VM instance {{(pid=62600) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 940.557671] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-fd2a2e22-128b-4f3f-a95e-1c3e2743ab8e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.563130] env[62600]: DEBUG nova.virt.hardware [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 940.563363] env[62600]: DEBUG nova.virt.hardware [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 940.563524] env[62600]: DEBUG nova.virt.hardware [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 940.563721] env[62600]: DEBUG nova.virt.hardware [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 940.563922] env[62600]: DEBUG nova.virt.hardware [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 940.564106] env[62600]: DEBUG nova.virt.hardware [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 940.564356] env[62600]: DEBUG nova.virt.hardware [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 940.564528] env[62600]: DEBUG nova.virt.hardware [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 940.564704] env[62600]: DEBUG nova.virt.hardware [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 940.564874] env[62600]: DEBUG nova.virt.hardware [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 940.565070] env[62600]: DEBUG nova.virt.hardware [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 940.565862] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ee46aa1-9c32-432b-a392-64adc13c3940 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.569328] env[62600]: DEBUG oslo_vmware.api [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for the task: (returnval){ [ 940.569328] env[62600]: value = "task-1222779" [ 940.569328] env[62600]: _type = "Task" [ 940.569328] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.576273] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65188745-f2ff-4dad-85c9-d6acdb63bb9b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.583192] env[62600]: DEBUG oslo_vmware.api [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222779, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.799366] env[62600]: DEBUG nova.scheduler.client.report [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 115 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 940.799539] env[62600]: DEBUG nova.compute.provider_tree [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 115 to 116 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 940.799781] env[62600]: DEBUG nova.compute.provider_tree [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 940.823205] env[62600]: DEBUG nova.compute.manager [req-9a0ae1e3-2ad4-4cab-9eff-4971565aba02 req-69deb4d8-33d3-4352-b1e0-93c6683e54f0 service nova] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Received event network-vif-plugged-64a8ba35-fe05-4c77-ac73-f420bd31a46d {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.825920] env[62600]: DEBUG oslo_concurrency.lockutils [req-9a0ae1e3-2ad4-4cab-9eff-4971565aba02 req-69deb4d8-33d3-4352-b1e0-93c6683e54f0 service nova] Acquiring lock "84dc284d-c7f4-4fc2-b539-8d4820429b02-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.825920] env[62600]: DEBUG oslo_concurrency.lockutils [req-9a0ae1e3-2ad4-4cab-9eff-4971565aba02 req-69deb4d8-33d3-4352-b1e0-93c6683e54f0 service nova] Lock "84dc284d-c7f4-4fc2-b539-8d4820429b02-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.825920] env[62600]: DEBUG oslo_concurrency.lockutils [req-9a0ae1e3-2ad4-4cab-9eff-4971565aba02 req-69deb4d8-33d3-4352-b1e0-93c6683e54f0 service nova] Lock "84dc284d-c7f4-4fc2-b539-8d4820429b02-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.825920] env[62600]: DEBUG nova.compute.manager [req-9a0ae1e3-2ad4-4cab-9eff-4971565aba02 req-69deb4d8-33d3-4352-b1e0-93c6683e54f0 service nova] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] No waiting events found dispatching network-vif-plugged-64a8ba35-fe05-4c77-ac73-f420bd31a46d {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 940.825920] env[62600]: WARNING nova.compute.manager [req-9a0ae1e3-2ad4-4cab-9eff-4971565aba02 req-69deb4d8-33d3-4352-b1e0-93c6683e54f0 service nova] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Received unexpected event network-vif-plugged-64a8ba35-fe05-4c77-ac73-f420bd31a46d for instance with vm_state building and task_state spawning. [ 940.902723] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222778, 'name': Rename_Task, 'duration_secs': 0.140404} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.903169] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 940.903381] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-100e75d0-addf-4804-ab5e-e28919aa6654 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.913686] env[62600]: DEBUG nova.network.neutron [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Successfully updated port: 64a8ba35-fe05-4c77-ac73-f420bd31a46d {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 940.915269] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 940.915269] env[62600]: value = "task-1222780" [ 940.915269] env[62600]: _type = "Task" [ 940.915269] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.923904] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222780, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.080443] env[62600]: DEBUG oslo_vmware.api [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222779, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.305967] env[62600]: DEBUG oslo_concurrency.lockutils [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.787s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.308409] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.206s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.308700] env[62600]: DEBUG nova.objects.instance [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lazy-loading 'resources' on Instance uuid 6447a26d-fc30-41eb-bd42-94341a86c2e8 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 941.325948] env[62600]: INFO nova.scheduler.client.report [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Deleted allocations for instance 9d899d96-9f4f-41d1-a368-3fde5efc110a [ 941.419745] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "refresh_cache-84dc284d-c7f4-4fc2-b539-8d4820429b02" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.419875] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquired lock "refresh_cache-84dc284d-c7f4-4fc2-b539-8d4820429b02" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.420045] env[62600]: DEBUG nova.network.neutron [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 941.427543] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222780, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.583443] env[62600]: DEBUG oslo_vmware.api [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222779, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.833365] env[62600]: DEBUG oslo_concurrency.lockutils [None req-578fe811-f413-4ac6-85eb-49c77c0f9497 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "9d899d96-9f4f-41d1-a368-3fde5efc110a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.861s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.927621] env[62600]: DEBUG oslo_vmware.api [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222780, 'name': PowerOnVM_Task, 'duration_secs': 0.862418} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.930315] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 941.930526] env[62600]: INFO nova.compute.manager [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Took 8.51 seconds to spawn the instance on the hypervisor. [ 941.930804] env[62600]: DEBUG nova.compute.manager [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 941.931806] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d28ac132-ff9b-40c5-951c-dad8b984156b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.957675] env[62600]: DEBUG nova.network.neutron [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 942.007854] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed5cc1a5-5847-4ac1-a3a4-33e67ed387fc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.016124] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e872ffb3-9266-46ba-a58b-ca3b7a4ea67f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.048799] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f281788-39c3-43c6-9dc1-cf6b372b502e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.056677] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a59e8040-cefd-49b7-b07f-ba1350332ca5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.070162] env[62600]: DEBUG nova.compute.provider_tree [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 942.081415] env[62600]: DEBUG oslo_vmware.api [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222779, 'name': RemoveSnapshot_Task, 'duration_secs': 1.11718} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.081679] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Deleted Snapshot of the VM instance {{(pid=62600) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 942.082045] env[62600]: INFO nova.compute.manager [None req-6cab91c5-8fb2-4344-a57c-d03eb66a6ccf tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Took 17.42 seconds to snapshot the instance on the hypervisor. [ 942.123506] env[62600]: DEBUG nova.network.neutron [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Updating instance_info_cache with network_info: [{"id": "64a8ba35-fe05-4c77-ac73-f420bd31a46d", "address": "fa:16:3e:ca:38:b8", "network": {"id": "f59f04dc-029f-42be-92de-1d7d3e973fd7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-36149453-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "734eda84fbe4493ba1f3db243a2bdb09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb971244-43ba-41b4-a6a2-a4558548012c", "external-id": "nsx-vlan-transportzone-873", "segmentation_id": 873, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64a8ba35-fe", "ovs_interfaceid": "64a8ba35-fe05-4c77-ac73-f420bd31a46d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.453098] env[62600]: INFO nova.compute.manager [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Took 26.25 seconds to build instance. [ 942.573081] env[62600]: DEBUG nova.scheduler.client.report [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 942.625990] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Releasing lock "refresh_cache-84dc284d-c7f4-4fc2-b539-8d4820429b02" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.626326] env[62600]: DEBUG nova.compute.manager [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Instance network_info: |[{"id": "64a8ba35-fe05-4c77-ac73-f420bd31a46d", "address": "fa:16:3e:ca:38:b8", "network": {"id": "f59f04dc-029f-42be-92de-1d7d3e973fd7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-36149453-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "734eda84fbe4493ba1f3db243a2bdb09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb971244-43ba-41b4-a6a2-a4558548012c", "external-id": "nsx-vlan-transportzone-873", "segmentation_id": 873, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64a8ba35-fe", "ovs_interfaceid": "64a8ba35-fe05-4c77-ac73-f420bd31a46d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 942.626766] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:38:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cb971244-43ba-41b4-a6a2-a4558548012c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '64a8ba35-fe05-4c77-ac73-f420bd31a46d', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 942.634253] env[62600]: DEBUG oslo.service.loopingcall [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.634477] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 942.634714] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0f86cd37-9136-40a6-8244-558963bd7741 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.658228] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 942.658228] env[62600]: value = "task-1222781" [ 942.658228] env[62600]: _type = "Task" [ 942.658228] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.666680] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222781, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.881584] env[62600]: DEBUG nova.compute.manager [req-bfd068db-e80a-4374-a3ed-7a72284d26ec req-e07f45e7-4b03-4498-8538-bcaf052a5124 service nova] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Received event network-changed-64a8ba35-fe05-4c77-ac73-f420bd31a46d {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.881802] env[62600]: DEBUG nova.compute.manager [req-bfd068db-e80a-4374-a3ed-7a72284d26ec req-e07f45e7-4b03-4498-8538-bcaf052a5124 service nova] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Refreshing instance network info cache due to event network-changed-64a8ba35-fe05-4c77-ac73-f420bd31a46d. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 942.882293] env[62600]: DEBUG oslo_concurrency.lockutils [req-bfd068db-e80a-4374-a3ed-7a72284d26ec req-e07f45e7-4b03-4498-8538-bcaf052a5124 service nova] Acquiring lock "refresh_cache-84dc284d-c7f4-4fc2-b539-8d4820429b02" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.882499] env[62600]: DEBUG oslo_concurrency.lockutils [req-bfd068db-e80a-4374-a3ed-7a72284d26ec req-e07f45e7-4b03-4498-8538-bcaf052a5124 service nova] Acquired lock "refresh_cache-84dc284d-c7f4-4fc2-b539-8d4820429b02" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.882726] env[62600]: DEBUG nova.network.neutron [req-bfd068db-e80a-4374-a3ed-7a72284d26ec req-e07f45e7-4b03-4498-8538-bcaf052a5124 service nova] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Refreshing network info cache for port 64a8ba35-fe05-4c77-ac73-f420bd31a46d {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 942.954838] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ccc08606-b686-4067-9b20-0c29cce93778 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "f18df958-2885-4d8c-a9ad-140faaed1178" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.762s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.078334] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.770s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.081564] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.976s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.082997] env[62600]: INFO nova.compute.claims [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 943.103093] env[62600]: INFO nova.scheduler.client.report [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleted allocations for instance 6447a26d-fc30-41eb-bd42-94341a86c2e8 [ 943.167890] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222781, 'name': CreateVM_Task, 'duration_secs': 0.346132} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.169025] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 943.169123] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.169308] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.169666] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 943.169922] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6fb98439-2885-4ee4-8c07-3a156a30bb5d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.175490] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 943.175490] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52dd6118-6ab1-09a0-95a2-9c76be0df994" [ 943.175490] env[62600]: _type = "Task" [ 943.175490] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.186967] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52dd6118-6ab1-09a0-95a2-9c76be0df994, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.189822] env[62600]: INFO nova.compute.manager [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Rebuilding instance [ 943.229767] env[62600]: DEBUG nova.compute.manager [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 943.230639] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-240536ad-ccf6-470a-b3b4-4ed0df2c3e8e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.581439] env[62600]: DEBUG nova.network.neutron [req-bfd068db-e80a-4374-a3ed-7a72284d26ec req-e07f45e7-4b03-4498-8538-bcaf052a5124 service nova] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Updated VIF entry in instance network info cache for port 64a8ba35-fe05-4c77-ac73-f420bd31a46d. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 943.581812] env[62600]: DEBUG nova.network.neutron [req-bfd068db-e80a-4374-a3ed-7a72284d26ec req-e07f45e7-4b03-4498-8538-bcaf052a5124 service nova] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Updating instance_info_cache with network_info: [{"id": "64a8ba35-fe05-4c77-ac73-f420bd31a46d", "address": "fa:16:3e:ca:38:b8", "network": {"id": "f59f04dc-029f-42be-92de-1d7d3e973fd7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-36149453-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "734eda84fbe4493ba1f3db243a2bdb09", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cb971244-43ba-41b4-a6a2-a4558548012c", "external-id": "nsx-vlan-transportzone-873", "segmentation_id": 873, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64a8ba35-fe", "ovs_interfaceid": "64a8ba35-fe05-4c77-ac73-f420bd31a46d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.614291] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fc9aa06b-af6e-439e-b721-67bfed5e7757 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "6447a26d-fc30-41eb-bd42-94341a86c2e8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.406s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.686430] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52dd6118-6ab1-09a0-95a2-9c76be0df994, 'name': SearchDatastore_Task, 'duration_secs': 0.010631} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.686768] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.687033] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 943.687287] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.687438] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.687625] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.690529] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4f03b2a0-7aa5-4e30-9f9f-fb0261f4df7b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.700994] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.701219] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 943.701987] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56eacd46-60bc-413d-8ad5-266c8080d4e9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.710105] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 943.710105] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f5a5cb-d69b-41e0-394d-3f58c0ff2ffe" [ 943.710105] env[62600]: _type = "Task" [ 943.710105] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.719520] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f5a5cb-d69b-41e0-394d-3f58c0ff2ffe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.743016] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 943.743363] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ac5f5f5c-bee8-4ca6-ac5d-68ac6c344ca8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.752638] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 943.752638] env[62600]: value = "task-1222782" [ 943.752638] env[62600]: _type = "Task" [ 943.752638] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.762054] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222782, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.085022] env[62600]: DEBUG oslo_concurrency.lockutils [req-bfd068db-e80a-4374-a3ed-7a72284d26ec req-e07f45e7-4b03-4498-8538-bcaf052a5124 service nova] Releasing lock "refresh_cache-84dc284d-c7f4-4fc2-b539-8d4820429b02" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.194313] env[62600]: DEBUG nova.compute.manager [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 944.195114] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff8b4abd-9e30-4d53-9012-fbe07d0b34ad {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.219840] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f5a5cb-d69b-41e0-394d-3f58c0ff2ffe, 'name': SearchDatastore_Task, 'duration_secs': 0.010874} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.224686] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e29e1111-4509-4137-ab3b-781700904de8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.231906] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 944.231906] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e6950c-6c03-1bc9-90c3-41b4516ada25" [ 944.231906] env[62600]: _type = "Task" [ 944.231906] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.244266] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e6950c-6c03-1bc9-90c3-41b4516ada25, 'name': SearchDatastore_Task, 'duration_secs': 0.010097} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.244668] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.244954] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 84dc284d-c7f4-4fc2-b539-8d4820429b02/84dc284d-c7f4-4fc2-b539-8d4820429b02.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 944.248382] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8d5b26bd-0c1d-4615-a1f9-aee50e8797f8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.258060] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 944.258060] env[62600]: value = "task-1222783" [ 944.258060] env[62600]: _type = "Task" [ 944.258060] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.267204] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222782, 'name': PowerOffVM_Task, 'duration_secs': 0.181429} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.268121] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 944.268370] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 944.269175] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1eda32d-4b3f-409d-a2bb-628000807638 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.274452] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222783, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.279722] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 944.279977] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1940e175-2229-49ab-a210-29d07e9eec7d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.287669] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7185287e-4113-4990-ae91-7c69a839936f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.296328] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89b4f0cd-fbf3-4f27-998c-7d80711d57e5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.326536] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84974053-8082-4b8e-bf75-cc87f92db382 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.334154] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f20ce05a-4236-4f39-a60a-3f20b3ef5ba2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.347677] env[62600]: DEBUG nova.compute.provider_tree [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 944.709497] env[62600]: INFO nova.compute.manager [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] instance snapshotting [ 944.713730] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b1641fc-e007-4e73-a85f-8d6570951655 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.739299] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e9e35e7-f12e-408f-8275-65ffcadc948a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.766845] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "4d12c133-84d7-4ce9-83db-e4b70b8947f3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.767181] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "4d12c133-84d7-4ce9-83db-e4b70b8947f3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.778151] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222783, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.780344] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 944.780589] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 944.780776] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleting the datastore file [datastore2] f18df958-2885-4d8c-a9ad-140faaed1178 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 944.781534] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-32edcf01-4057-4133-a49d-a63eea9d9eb2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.789843] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 944.789843] env[62600]: value = "task-1222785" [ 944.789843] env[62600]: _type = "Task" [ 944.789843] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.798120] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222785, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.850963] env[62600]: DEBUG nova.scheduler.client.report [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 945.251769] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Creating Snapshot of the VM instance {{(pid=62600) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 945.252250] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a50336ca-c9c6-4dbf-9a28-43998210edb0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.260489] env[62600]: DEBUG oslo_vmware.api [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for the task: (returnval){ [ 945.260489] env[62600]: value = "task-1222786" [ 945.260489] env[62600]: _type = "Task" [ 945.260489] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.270233] env[62600]: DEBUG nova.compute.manager [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 945.279031] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222783, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.551392} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.279031] env[62600]: DEBUG oslo_vmware.api [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222786, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.279479] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 84dc284d-c7f4-4fc2-b539-8d4820429b02/84dc284d-c7f4-4fc2-b539-8d4820429b02.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 945.280309] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 945.280678] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6636e419-62a7-4c19-b1cb-a45635e1d9e0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.288932] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 945.288932] env[62600]: value = "task-1222787" [ 945.288932] env[62600]: _type = "Task" [ 945.288932] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.303484] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222787, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.308206] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222785, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164535} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.309177] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 945.309396] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 945.309602] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 945.356869] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.276s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.357475] env[62600]: DEBUG nova.compute.manager [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 945.359931] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.200s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.361387] env[62600]: INFO nova.compute.claims [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 945.774678] env[62600]: DEBUG oslo_vmware.api [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222786, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.800076] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222787, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063092} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.801228] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.801566] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 945.802382] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a2080e7-4434-4582-8bab-3727527d1d84 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.828221] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 84dc284d-c7f4-4fc2-b539-8d4820429b02/84dc284d-c7f4-4fc2-b539-8d4820429b02.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.828752] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27c39de2-b0da-46a7-80cf-78c92718c4d2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.852282] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 945.852282] env[62600]: value = "task-1222788" [ 945.852282] env[62600]: _type = "Task" [ 945.852282] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.863333] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222788, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.865896] env[62600]: DEBUG nova.compute.utils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 945.869655] env[62600]: DEBUG nova.compute.manager [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 945.869824] env[62600]: DEBUG nova.network.neutron [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 945.912309] env[62600]: DEBUG nova.policy [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '41f578fc709a497abd53d1a4f0540f6d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5e7eca5e197e47759eeea423e280986f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 946.148393] env[62600]: DEBUG nova.network.neutron [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Successfully created port: 2bf1b736-051a-4ca6-8800-a34547f63394 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 946.277195] env[62600]: DEBUG oslo_vmware.api [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222786, 'name': CreateSnapshot_Task, 'duration_secs': 0.80625} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.277527] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Created Snapshot of the VM instance {{(pid=62600) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 946.278312] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-160ad393-c887-4666-9b18-dd71f82583c9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.351865] env[62600]: DEBUG nova.virt.hardware [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 946.352146] env[62600]: DEBUG nova.virt.hardware [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 946.352314] env[62600]: DEBUG nova.virt.hardware [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 946.352508] env[62600]: DEBUG nova.virt.hardware [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 946.352669] env[62600]: DEBUG nova.virt.hardware [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 946.352971] env[62600]: DEBUG nova.virt.hardware [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 946.353194] env[62600]: DEBUG nova.virt.hardware [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 946.353321] env[62600]: DEBUG nova.virt.hardware [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 946.353452] env[62600]: DEBUG nova.virt.hardware [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 946.353649] env[62600]: DEBUG nova.virt.hardware [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 946.353847] env[62600]: DEBUG nova.virt.hardware [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 946.354877] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a885f61-466a-4f42-b629-264859ccc9ae {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.366589] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222788, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.369634] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81dc5757-9189-4f68-a1ee-b6caef3e2197 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.373648] env[62600]: DEBUG nova.compute.manager [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 946.392256] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:fa:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43798f54-0c07-4417-a23f-58bb6b7e204b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '874235d5-d1b2-4382-a30a-2e187141de5b', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 946.400417] env[62600]: DEBUG oslo.service.loopingcall [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.404179] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 946.404595] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-749a0437-62d2-41d4-9d26-b4c4c3d4a25d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.428885] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 946.428885] env[62600]: value = "task-1222789" [ 946.428885] env[62600]: _type = "Task" [ 946.428885] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.438456] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222789, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.595773] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d1560f-932d-4147-b568-94b648e9df7e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.603958] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6363a637-1fc4-49b7-be79-c60fcffefd8f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.635108] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a32eb6-658a-411a-af67-593a28ff7fd8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.642937] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43cd9375-403f-457b-b1db-d8fbbe109b08 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.656346] env[62600]: DEBUG nova.compute.provider_tree [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.797658] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Creating linked-clone VM from snapshot {{(pid=62600) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 946.798900] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-2b61c1f7-6ec1-4485-af62-7e5eacad68b4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.809676] env[62600]: DEBUG oslo_vmware.api [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for the task: (returnval){ [ 946.809676] env[62600]: value = "task-1222790" [ 946.809676] env[62600]: _type = "Task" [ 946.809676] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.819681] env[62600]: DEBUG oslo_vmware.api [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222790, 'name': CloneVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.864041] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222788, 'name': ReconfigVM_Task, 'duration_secs': 0.556728} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.865469] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 84dc284d-c7f4-4fc2-b539-8d4820429b02/84dc284d-c7f4-4fc2-b539-8d4820429b02.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.865469] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2edb386f-53c3-4059-8697-1b167bffb6ab {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.873828] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 946.873828] env[62600]: value = "task-1222791" [ 946.873828] env[62600]: _type = "Task" [ 946.873828] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.890823] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222791, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.940036] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222789, 'name': CreateVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.159706] env[62600]: DEBUG nova.scheduler.client.report [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 947.319515] env[62600]: DEBUG oslo_vmware.api [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222790, 'name': CloneVM_Task} progress is 94%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.384978] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222791, 'name': Rename_Task, 'duration_secs': 0.156373} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.384978] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 947.385240] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2276137a-f18c-49aa-809f-d08468bab31a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.390166] env[62600]: DEBUG nova.compute.manager [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 947.393658] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 947.393658] env[62600]: value = "task-1222792" [ 947.393658] env[62600]: _type = "Task" [ 947.393658] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.402154] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222792, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.411835] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 947.412056] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 947.412232] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 947.412419] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 947.412572] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 947.412728] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 947.412960] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 947.413172] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 947.413358] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 947.413533] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 947.413717] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 947.414688] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214da02d-1404-42e4-a21b-a158012ab88c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.423147] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98bcdf9b-68d3-4120-879c-75ee0751673f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.450013] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222789, 'name': CreateVM_Task, 'duration_secs': 0.513158} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.450221] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 947.450988] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.451198] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.451533] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 947.451801] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-696a96c5-3fe6-4931-85d5-25e376774430 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.458965] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 947.458965] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52006411-1c04-acfa-e8ff-998fa27637cd" [ 947.458965] env[62600]: _type = "Task" [ 947.458965] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.468414] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52006411-1c04-acfa-e8ff-998fa27637cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.555969] env[62600]: DEBUG nova.compute.manager [req-65b4c17a-5ed3-4575-84f6-f1941c6de6c7 req-44f8565f-e61e-4c5e-9917-cb345be9a779 service nova] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Received event network-vif-plugged-2bf1b736-051a-4ca6-8800-a34547f63394 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.556308] env[62600]: DEBUG oslo_concurrency.lockutils [req-65b4c17a-5ed3-4575-84f6-f1941c6de6c7 req-44f8565f-e61e-4c5e-9917-cb345be9a779 service nova] Acquiring lock "784e6d19-36ef-4c01-9e9f-5a083ed08608-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.556745] env[62600]: DEBUG oslo_concurrency.lockutils [req-65b4c17a-5ed3-4575-84f6-f1941c6de6c7 req-44f8565f-e61e-4c5e-9917-cb345be9a779 service nova] Lock "784e6d19-36ef-4c01-9e9f-5a083ed08608-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.556745] env[62600]: DEBUG oslo_concurrency.lockutils [req-65b4c17a-5ed3-4575-84f6-f1941c6de6c7 req-44f8565f-e61e-4c5e-9917-cb345be9a779 service nova] Lock "784e6d19-36ef-4c01-9e9f-5a083ed08608-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.556829] env[62600]: DEBUG nova.compute.manager [req-65b4c17a-5ed3-4575-84f6-f1941c6de6c7 req-44f8565f-e61e-4c5e-9917-cb345be9a779 service nova] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] No waiting events found dispatching network-vif-plugged-2bf1b736-051a-4ca6-8800-a34547f63394 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 947.556992] env[62600]: WARNING nova.compute.manager [req-65b4c17a-5ed3-4575-84f6-f1941c6de6c7 req-44f8565f-e61e-4c5e-9917-cb345be9a779 service nova] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Received unexpected event network-vif-plugged-2bf1b736-051a-4ca6-8800-a34547f63394 for instance with vm_state building and task_state spawning. [ 947.647411] env[62600]: DEBUG nova.network.neutron [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Successfully updated port: 2bf1b736-051a-4ca6-8800-a34547f63394 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 947.665252] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.305s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.665871] env[62600]: DEBUG nova.compute.manager [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 947.669596] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.399s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.669884] env[62600]: DEBUG nova.objects.instance [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lazy-loading 'resources' on Instance uuid afe838cc-e086-4986-87ec-4e1266bcaf60 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 947.822108] env[62600]: DEBUG oslo_vmware.api [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222790, 'name': CloneVM_Task} progress is 95%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.903836] env[62600]: DEBUG oslo_vmware.api [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222792, 'name': PowerOnVM_Task, 'duration_secs': 0.470923} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.904210] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 947.904463] env[62600]: INFO nova.compute.manager [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Took 7.36 seconds to spawn the instance on the hypervisor. [ 947.904688] env[62600]: DEBUG nova.compute.manager [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 947.905534] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56997b6c-92a6-4b33-a1b1-09c7fc32b577 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.969879] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52006411-1c04-acfa-e8ff-998fa27637cd, 'name': SearchDatastore_Task, 'duration_secs': 0.014174} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.970209] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.970448] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 947.970690] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.970843] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.971051] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 947.971339] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cd472278-26db-4709-8070-a33cadada80d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.984813] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 947.985191] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 947.986157] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6da17b09-65a3-4ead-a525-35e16c7f3b09 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.993465] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 947.993465] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52730235-3ae3-a762-c4ce-e117dd3b7530" [ 947.993465] env[62600]: _type = "Task" [ 947.993465] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.002560] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52730235-3ae3-a762-c4ce-e117dd3b7530, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.153219] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "refresh_cache-784e6d19-36ef-4c01-9e9f-5a083ed08608" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.153434] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquired lock "refresh_cache-784e6d19-36ef-4c01-9e9f-5a083ed08608" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.153604] env[62600]: DEBUG nova.network.neutron [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 948.171458] env[62600]: DEBUG nova.compute.utils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 948.173199] env[62600]: DEBUG nova.compute.manager [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 948.173418] env[62600]: DEBUG nova.network.neutron [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 948.215419] env[62600]: DEBUG nova.policy [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '41f578fc709a497abd53d1a4f0540f6d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5e7eca5e197e47759eeea423e280986f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 948.323575] env[62600]: DEBUG oslo_vmware.api [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222790, 'name': CloneVM_Task, 'duration_secs': 1.333077} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.326076] env[62600]: INFO nova.virt.vmwareapi.vmops [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Created linked-clone VM from snapshot [ 948.327181] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06d903a6-291f-4435-bdaa-6c0ec370248d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.334817] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Uploading image bc24c6fc-64cd-481c-956d-7fa0f16f714f {{(pid=62600) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 948.359202] env[62600]: DEBUG oslo_vmware.rw_handles [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 948.359202] env[62600]: value = "vm-264342" [ 948.359202] env[62600]: _type = "VirtualMachine" [ 948.359202] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 948.359489] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-40549020-0579-4ec3-8eec-dfc96aaf0499 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.369631] env[62600]: DEBUG oslo_vmware.rw_handles [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Lease: (returnval){ [ 948.369631] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525dab1c-34e4-0f1b-219b-d1296b974fe2" [ 948.369631] env[62600]: _type = "HttpNfcLease" [ 948.369631] env[62600]: } obtained for exporting VM: (result){ [ 948.369631] env[62600]: value = "vm-264342" [ 948.369631] env[62600]: _type = "VirtualMachine" [ 948.369631] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 948.369904] env[62600]: DEBUG oslo_vmware.api [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for the lease: (returnval){ [ 948.369904] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525dab1c-34e4-0f1b-219b-d1296b974fe2" [ 948.369904] env[62600]: _type = "HttpNfcLease" [ 948.369904] env[62600]: } to be ready. {{(pid=62600) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 948.374649] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d0200f-ad47-4ac8-b6b9-130667e582c3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.380739] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 948.380739] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525dab1c-34e4-0f1b-219b-d1296b974fe2" [ 948.380739] env[62600]: _type = "HttpNfcLease" [ 948.380739] env[62600]: } is initializing. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 948.383634] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9975f475-b91b-46af-84f5-cff0e56ffb02 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.413125] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57a6162d-d8de-4a9a-9926-88d14e72030b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.428495] env[62600]: INFO nova.compute.manager [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Took 25.59 seconds to build instance. [ 948.430667] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b71a60-17a4-4470-9d29-bc1872202088 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.446926] env[62600]: DEBUG nova.compute.provider_tree [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.505838] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52730235-3ae3-a762-c4ce-e117dd3b7530, 'name': SearchDatastore_Task, 'duration_secs': 0.010391} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.506363] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb03c21f-c8df-4540-9f9b-deccaa500996 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.511297] env[62600]: DEBUG nova.network.neutron [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Successfully created port: df39b522-a9a3-4d74-afe8-c68d5355ee10 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 948.514778] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 948.514778] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]523e18fb-8545-0a11-6686-a89a0fd65fb0" [ 948.514778] env[62600]: _type = "Task" [ 948.514778] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.523672] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]523e18fb-8545-0a11-6686-a89a0fd65fb0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.678802] env[62600]: DEBUG nova.compute.manager [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 948.699519] env[62600]: DEBUG nova.network.neutron [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 948.874936] env[62600]: DEBUG nova.network.neutron [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Updating instance_info_cache with network_info: [{"id": "2bf1b736-051a-4ca6-8800-a34547f63394", "address": "fa:16:3e:52:2a:aa", "network": {"id": "353f7a99-4652-4a0d-be26-539c5c42b3d2", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1505004576-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7eca5e197e47759eeea423e280986f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bf1b736-05", "ovs_interfaceid": "2bf1b736-051a-4ca6-8800-a34547f63394", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.880280] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 948.880280] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525dab1c-34e4-0f1b-219b-d1296b974fe2" [ 948.880280] env[62600]: _type = "HttpNfcLease" [ 948.880280] env[62600]: } is ready. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 948.880594] env[62600]: DEBUG oslo_vmware.rw_handles [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 948.880594] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]525dab1c-34e4-0f1b-219b-d1296b974fe2" [ 948.880594] env[62600]: _type = "HttpNfcLease" [ 948.880594] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 948.881378] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84279ddb-6e0c-49d2-aece-50a1e17b99b9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.889466] env[62600]: DEBUG oslo_vmware.rw_handles [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f06b44-4035-1d45-71c8-ffa1a25bc384/disk-0.vmdk from lease info. {{(pid=62600) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 948.889649] env[62600]: DEBUG oslo_vmware.rw_handles [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f06b44-4035-1d45-71c8-ffa1a25bc384/disk-0.vmdk for reading. {{(pid=62600) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 948.946214] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df79b653-6331-4854-b9fd-848803778079 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "84dc284d-c7f4-4fc2-b539-8d4820429b02" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.122s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.950502] env[62600]: DEBUG nova.scheduler.client.report [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 948.983697] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-661a6f51-b5ea-4611-9e49-95c71932548c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.033335] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]523e18fb-8545-0a11-6686-a89a0fd65fb0, 'name': SearchDatastore_Task, 'duration_secs': 0.011043} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.033335] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.033491] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] f18df958-2885-4d8c-a9ad-140faaed1178/f18df958-2885-4d8c-a9ad-140faaed1178.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 949.034306] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4a1963ce-c6e9-4f60-bd3f-f120d7ae3ad6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.043500] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 949.043500] env[62600]: value = "task-1222794" [ 949.043500] env[62600]: _type = "Task" [ 949.043500] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.052711] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222794, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.377801] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Releasing lock "refresh_cache-784e6d19-36ef-4c01-9e9f-5a083ed08608" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.379977] env[62600]: DEBUG nova.compute.manager [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Instance network_info: |[{"id": "2bf1b736-051a-4ca6-8800-a34547f63394", "address": "fa:16:3e:52:2a:aa", "network": {"id": "353f7a99-4652-4a0d-be26-539c5c42b3d2", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1505004576-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7eca5e197e47759eeea423e280986f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bf1b736-05", "ovs_interfaceid": "2bf1b736-051a-4ca6-8800-a34547f63394", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 949.379977] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:2a:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd998416-f3d6-4a62-b828-5011063ce76a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2bf1b736-051a-4ca6-8800-a34547f63394', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 949.387949] env[62600]: DEBUG oslo.service.loopingcall [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 949.388355] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 949.388598] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ee6e89b0-18a5-425c-9e94-7db8adce3c62 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.411647] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 949.411647] env[62600]: value = "task-1222795" [ 949.411647] env[62600]: _type = "Task" [ 949.411647] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.421676] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222795, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.456204] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.786s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.459994] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.248s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.462174] env[62600]: INFO nova.compute.claims [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 949.490242] env[62600]: INFO nova.scheduler.client.report [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Deleted allocations for instance afe838cc-e086-4986-87ec-4e1266bcaf60 [ 949.555915] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222794, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474188} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.557341] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] f18df958-2885-4d8c-a9ad-140faaed1178/f18df958-2885-4d8c-a9ad-140faaed1178.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 949.557341] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 949.557744] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3903b39a-f161-4ca5-b754-bacfd5c0f447 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.574344] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 949.574344] env[62600]: value = "task-1222796" [ 949.574344] env[62600]: _type = "Task" [ 949.574344] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.585847] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222796, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.589883] env[62600]: DEBUG nova.compute.manager [req-4309d76b-c87d-4028-8541-df5518dcdb5b req-9abe8835-5c0d-4db1-9f18-a4eded5a5bf0 service nova] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Received event network-changed-2bf1b736-051a-4ca6-8800-a34547f63394 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.590494] env[62600]: DEBUG nova.compute.manager [req-4309d76b-c87d-4028-8541-df5518dcdb5b req-9abe8835-5c0d-4db1-9f18-a4eded5a5bf0 service nova] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Refreshing instance network info cache due to event network-changed-2bf1b736-051a-4ca6-8800-a34547f63394. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 949.590494] env[62600]: DEBUG oslo_concurrency.lockutils [req-4309d76b-c87d-4028-8541-df5518dcdb5b req-9abe8835-5c0d-4db1-9f18-a4eded5a5bf0 service nova] Acquiring lock "refresh_cache-784e6d19-36ef-4c01-9e9f-5a083ed08608" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.590660] env[62600]: DEBUG oslo_concurrency.lockutils [req-4309d76b-c87d-4028-8541-df5518dcdb5b req-9abe8835-5c0d-4db1-9f18-a4eded5a5bf0 service nova] Acquired lock "refresh_cache-784e6d19-36ef-4c01-9e9f-5a083ed08608" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.591237] env[62600]: DEBUG nova.network.neutron [req-4309d76b-c87d-4028-8541-df5518dcdb5b req-9abe8835-5c0d-4db1-9f18-a4eded5a5bf0 service nova] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Refreshing network info cache for port 2bf1b736-051a-4ca6-8800-a34547f63394 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 949.688938] env[62600]: DEBUG nova.compute.manager [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 949.767305] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 949.768308] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 949.768308] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 949.768308] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 949.768308] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 949.768517] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 949.768636] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 949.768817] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 949.769038] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 949.769252] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 949.769447] env[62600]: DEBUG nova.virt.hardware [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 949.770369] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee50a5d0-788b-42d4-86ee-8bec8172ee89 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.779518] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc32004a-5ca5-4400-8d77-2a86ffca5590 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.923527] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222795, 'name': CreateVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.001023] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d590c1dd-7aaf-4ecb-97b6-f2f5e7b5eebb tempest-VolumesAdminNegativeTest-984915564 tempest-VolumesAdminNegativeTest-984915564-project-member] Lock "afe838cc-e086-4986-87ec-4e1266bcaf60" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.932s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.090644] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222796, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070248} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.090812] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 950.091702] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccf0529a-bc4c-4c16-b6af-823609d025e8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.124095] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] f18df958-2885-4d8c-a9ad-140faaed1178/f18df958-2885-4d8c-a9ad-140faaed1178.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 950.125534] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-859b5e35-0961-420e-84a8-9c8fb536b742 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.145304] env[62600]: DEBUG nova.network.neutron [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Successfully updated port: df39b522-a9a3-4d74-afe8-c68d5355ee10 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 950.156124] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 950.156124] env[62600]: value = "task-1222797" [ 950.156124] env[62600]: _type = "Task" [ 950.156124] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.170036] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222797, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.331290] env[62600]: DEBUG nova.network.neutron [req-4309d76b-c87d-4028-8541-df5518dcdb5b req-9abe8835-5c0d-4db1-9f18-a4eded5a5bf0 service nova] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Updated VIF entry in instance network info cache for port 2bf1b736-051a-4ca6-8800-a34547f63394. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 950.331869] env[62600]: DEBUG nova.network.neutron [req-4309d76b-c87d-4028-8541-df5518dcdb5b req-9abe8835-5c0d-4db1-9f18-a4eded5a5bf0 service nova] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Updating instance_info_cache with network_info: [{"id": "2bf1b736-051a-4ca6-8800-a34547f63394", "address": "fa:16:3e:52:2a:aa", "network": {"id": "353f7a99-4652-4a0d-be26-539c5c42b3d2", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1505004576-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7eca5e197e47759eeea423e280986f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2bf1b736-05", "ovs_interfaceid": "2bf1b736-051a-4ca6-8800-a34547f63394", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.423348] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222795, 'name': CreateVM_Task, 'duration_secs': 0.615528} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.423668] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 950.424383] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.424561] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.425052] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 950.425379] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b815ccd5-1f47-43ef-ad12-0aec769f1ba3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.431336] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 950.431336] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528935c0-8bfd-de53-d33f-be7d27565559" [ 950.431336] env[62600]: _type = "Task" [ 950.431336] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.445390] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528935c0-8bfd-de53-d33f-be7d27565559, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.584116] env[62600]: DEBUG nova.compute.manager [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 950.587737] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c1bfac-55a5-4dba-b8f9-c1829c4abfab {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.651120] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "refresh_cache-81333628-2616-4557-88d1-0e7164e1b16d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.651329] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquired lock "refresh_cache-81333628-2616-4557-88d1-0e7164e1b16d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.651482] env[62600]: DEBUG nova.network.neutron [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 950.671169] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222797, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.724067] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88aaafc2-eb45-4d5a-bff3-53a5eac9bf61 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.733528] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d1061c8-055e-491d-b0fd-161f2968095b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.765619] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa0c5c8-15c4-489d-8680-b828c1f29f01 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.777324] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7140f8ec-a27e-4c67-b8c7-2fd9caf305b8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.788855] env[62600]: DEBUG nova.compute.provider_tree [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.834737] env[62600]: DEBUG oslo_concurrency.lockutils [req-4309d76b-c87d-4028-8541-df5518dcdb5b req-9abe8835-5c0d-4db1-9f18-a4eded5a5bf0 service nova] Releasing lock "refresh_cache-784e6d19-36ef-4c01-9e9f-5a083ed08608" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.945199] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528935c0-8bfd-de53-d33f-be7d27565559, 'name': SearchDatastore_Task, 'duration_secs': 0.013798} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.945199] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.945199] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 950.945199] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.945199] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.945199] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 950.945529] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5fe5aaa6-ae51-4927-813e-27ee374871b9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.955107] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 950.957706] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 950.957706] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be291e40-6215-4601-9472-1b729bb34aa7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.962748] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 950.962748] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5278d9e1-82d9-b4fb-9899-f759aab96696" [ 950.962748] env[62600]: _type = "Task" [ 950.962748] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.972802] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5278d9e1-82d9-b4fb-9899-f759aab96696, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.102296] env[62600]: INFO nova.compute.manager [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] instance snapshotting [ 951.106350] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2612a260-295e-41e9-8b45-60e4a562975e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.131633] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9c5e39-1fa5-4f21-9592-2dc6e653fe5b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.168319] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222797, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.185110] env[62600]: DEBUG nova.network.neutron [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 951.291822] env[62600]: DEBUG nova.scheduler.client.report [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 951.426166] env[62600]: DEBUG nova.network.neutron [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Updating instance_info_cache with network_info: [{"id": "df39b522-a9a3-4d74-afe8-c68d5355ee10", "address": "fa:16:3e:30:e5:f2", "network": {"id": "353f7a99-4652-4a0d-be26-539c5c42b3d2", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1505004576-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7eca5e197e47759eeea423e280986f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf39b522-a9", "ovs_interfaceid": "df39b522-a9a3-4d74-afe8-c68d5355ee10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.476682] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5278d9e1-82d9-b4fb-9899-f759aab96696, 'name': SearchDatastore_Task, 'duration_secs': 0.009381} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.477251] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1611da0d-dc62-480d-b1ce-9fd88de01b90 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.485487] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 951.485487] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c55768-731e-0915-c871-5ab047381c87" [ 951.485487] env[62600]: _type = "Task" [ 951.485487] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.497344] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c55768-731e-0915-c871-5ab047381c87, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.643965] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Creating Snapshot of the VM instance {{(pid=62600) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 951.644377] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a4f1dc1d-e7bf-4bb0-b499-7d5c2a46d919 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.652116] env[62600]: DEBUG nova.compute.manager [req-cda91479-dd82-4688-a86c-2aa4c0ae02df req-758c1c53-4554-4092-bdd2-39cc3d55c8a1 service nova] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Received event network-vif-plugged-df39b522-a9a3-4d74-afe8-c68d5355ee10 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.652832] env[62600]: DEBUG oslo_concurrency.lockutils [req-cda91479-dd82-4688-a86c-2aa4c0ae02df req-758c1c53-4554-4092-bdd2-39cc3d55c8a1 service nova] Acquiring lock "81333628-2616-4557-88d1-0e7164e1b16d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.653036] env[62600]: DEBUG oslo_concurrency.lockutils [req-cda91479-dd82-4688-a86c-2aa4c0ae02df req-758c1c53-4554-4092-bdd2-39cc3d55c8a1 service nova] Lock "81333628-2616-4557-88d1-0e7164e1b16d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.653255] env[62600]: DEBUG oslo_concurrency.lockutils [req-cda91479-dd82-4688-a86c-2aa4c0ae02df req-758c1c53-4554-4092-bdd2-39cc3d55c8a1 service nova] Lock "81333628-2616-4557-88d1-0e7164e1b16d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.653464] env[62600]: DEBUG nova.compute.manager [req-cda91479-dd82-4688-a86c-2aa4c0ae02df req-758c1c53-4554-4092-bdd2-39cc3d55c8a1 service nova] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] No waiting events found dispatching network-vif-plugged-df39b522-a9a3-4d74-afe8-c68d5355ee10 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 951.653675] env[62600]: WARNING nova.compute.manager [req-cda91479-dd82-4688-a86c-2aa4c0ae02df req-758c1c53-4554-4092-bdd2-39cc3d55c8a1 service nova] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Received unexpected event network-vif-plugged-df39b522-a9a3-4d74-afe8-c68d5355ee10 for instance with vm_state building and task_state spawning. [ 951.653918] env[62600]: DEBUG nova.compute.manager [req-cda91479-dd82-4688-a86c-2aa4c0ae02df req-758c1c53-4554-4092-bdd2-39cc3d55c8a1 service nova] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Received event network-changed-df39b522-a9a3-4d74-afe8-c68d5355ee10 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.654092] env[62600]: DEBUG nova.compute.manager [req-cda91479-dd82-4688-a86c-2aa4c0ae02df req-758c1c53-4554-4092-bdd2-39cc3d55c8a1 service nova] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Refreshing instance network info cache due to event network-changed-df39b522-a9a3-4d74-afe8-c68d5355ee10. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 951.654327] env[62600]: DEBUG oslo_concurrency.lockutils [req-cda91479-dd82-4688-a86c-2aa4c0ae02df req-758c1c53-4554-4092-bdd2-39cc3d55c8a1 service nova] Acquiring lock "refresh_cache-81333628-2616-4557-88d1-0e7164e1b16d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.656393] env[62600]: DEBUG oslo_vmware.api [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 951.656393] env[62600]: value = "task-1222798" [ 951.656393] env[62600]: _type = "Task" [ 951.656393] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.673710] env[62600]: DEBUG oslo_vmware.api [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222798, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.677158] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222797, 'name': ReconfigVM_Task, 'duration_secs': 1.114623} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.678073] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Reconfigured VM instance instance-00000053 to attach disk [datastore2] f18df958-2885-4d8c-a9ad-140faaed1178/f18df958-2885-4d8c-a9ad-140faaed1178.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 951.678510] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-013b55f4-890b-4fd1-aa95-0d14aa4e29a1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.687333] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 951.687333] env[62600]: value = "task-1222799" [ 951.687333] env[62600]: _type = "Task" [ 951.687333] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.699326] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222799, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.798935] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.338s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.798935] env[62600]: DEBUG nova.compute.manager [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 951.801607] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.795s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.803243] env[62600]: INFO nova.compute.claims [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 951.930823] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Releasing lock "refresh_cache-81333628-2616-4557-88d1-0e7164e1b16d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.931028] env[62600]: DEBUG nova.compute.manager [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Instance network_info: |[{"id": "df39b522-a9a3-4d74-afe8-c68d5355ee10", "address": "fa:16:3e:30:e5:f2", "network": {"id": "353f7a99-4652-4a0d-be26-539c5c42b3d2", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1505004576-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7eca5e197e47759eeea423e280986f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf39b522-a9", "ovs_interfaceid": "df39b522-a9a3-4d74-afe8-c68d5355ee10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 951.931384] env[62600]: DEBUG oslo_concurrency.lockutils [req-cda91479-dd82-4688-a86c-2aa4c0ae02df req-758c1c53-4554-4092-bdd2-39cc3d55c8a1 service nova] Acquired lock "refresh_cache-81333628-2616-4557-88d1-0e7164e1b16d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.931576] env[62600]: DEBUG nova.network.neutron [req-cda91479-dd82-4688-a86c-2aa4c0ae02df req-758c1c53-4554-4092-bdd2-39cc3d55c8a1 service nova] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Refreshing network info cache for port df39b522-a9a3-4d74-afe8-c68d5355ee10 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 951.932919] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:e5:f2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd998416-f3d6-4a62-b828-5011063ce76a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'df39b522-a9a3-4d74-afe8-c68d5355ee10', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 951.945538] env[62600]: DEBUG oslo.service.loopingcall [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.949927] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 951.950277] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-84cc6bd5-dfe4-4956-b315-47018953deee {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.978080] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 951.978080] env[62600]: value = "task-1222800" [ 951.978080] env[62600]: _type = "Task" [ 951.978080] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.987638] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222800, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.000633] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c55768-731e-0915-c871-5ab047381c87, 'name': SearchDatastore_Task, 'duration_secs': 0.010211} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.001183] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.001774] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 784e6d19-36ef-4c01-9e9f-5a083ed08608/784e6d19-36ef-4c01-9e9f-5a083ed08608.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 952.004019] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5705e0d4-52f6-4da4-a57d-2be3fac4b0be {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.014039] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 952.014039] env[62600]: value = "task-1222801" [ 952.014039] env[62600]: _type = "Task" [ 952.014039] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.027082] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222801, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.167739] env[62600]: DEBUG oslo_vmware.api [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222798, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.198579] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222799, 'name': Rename_Task, 'duration_secs': 0.167425} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.198915] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 952.199186] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f9c9a13e-82b4-48f8-80b0-6e96a390b36f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.207664] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 952.207664] env[62600]: value = "task-1222802" [ 952.207664] env[62600]: _type = "Task" [ 952.207664] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.218078] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222802, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.219087] env[62600]: DEBUG nova.network.neutron [req-cda91479-dd82-4688-a86c-2aa4c0ae02df req-758c1c53-4554-4092-bdd2-39cc3d55c8a1 service nova] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Updated VIF entry in instance network info cache for port df39b522-a9a3-4d74-afe8-c68d5355ee10. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 952.219456] env[62600]: DEBUG nova.network.neutron [req-cda91479-dd82-4688-a86c-2aa4c0ae02df req-758c1c53-4554-4092-bdd2-39cc3d55c8a1 service nova] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Updating instance_info_cache with network_info: [{"id": "df39b522-a9a3-4d74-afe8-c68d5355ee10", "address": "fa:16:3e:30:e5:f2", "network": {"id": "353f7a99-4652-4a0d-be26-539c5c42b3d2", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1505004576-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5e7eca5e197e47759eeea423e280986f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd998416-f3d6-4a62-b828-5011063ce76a", "external-id": "nsx-vlan-transportzone-57", "segmentation_id": 57, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdf39b522-a9", "ovs_interfaceid": "df39b522-a9a3-4d74-afe8-c68d5355ee10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.308230] env[62600]: DEBUG nova.compute.utils [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 952.313336] env[62600]: DEBUG nova.compute.manager [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 952.313520] env[62600]: DEBUG nova.network.neutron [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 952.396125] env[62600]: DEBUG nova.policy [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34319d68370c48ef93fa24a9d9132407', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bab74e3ae78248909dbcd483abbb8da2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 952.489480] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222800, 'name': CreateVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.525357] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222801, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.668447] env[62600]: DEBUG oslo_vmware.api [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222798, 'name': CreateSnapshot_Task, 'duration_secs': 0.750319} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.668953] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Created Snapshot of the VM instance {{(pid=62600) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 952.670061] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e579f296-9462-4a01-9013-25c1f6d1282d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.704172] env[62600]: DEBUG nova.network.neutron [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Successfully created port: 80551113-7315-4224-ac50-2f6b14d0a6b0 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 952.719021] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222802, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.721907] env[62600]: DEBUG oslo_concurrency.lockutils [req-cda91479-dd82-4688-a86c-2aa4c0ae02df req-758c1c53-4554-4092-bdd2-39cc3d55c8a1 service nova] Releasing lock "refresh_cache-81333628-2616-4557-88d1-0e7164e1b16d" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.814255] env[62600]: DEBUG nova.compute.manager [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 952.992729] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222800, 'name': CreateVM_Task, 'duration_secs': 0.526542} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.993049] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 952.993646] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.993824] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.994349] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 952.994645] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-608b7372-e36e-4f99-b2e5-160303b9f8c3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.003250] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 953.003250] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ecff50-a4c4-da54-5e5c-b8a63e33c5d4" [ 953.003250] env[62600]: _type = "Task" [ 953.003250] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.014546] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ecff50-a4c4-da54-5e5c-b8a63e33c5d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.026797] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222801, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.54533} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.027113] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 784e6d19-36ef-4c01-9e9f-5a083ed08608/784e6d19-36ef-4c01-9e9f-5a083ed08608.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 953.027359] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 953.027752] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0114c261-c5a5-48af-a35a-dadda68c3b28 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.038191] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 953.038191] env[62600]: value = "task-1222803" [ 953.038191] env[62600]: _type = "Task" [ 953.038191] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.052994] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222803, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.098742] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-769c442f-4c05-4700-b96b-866dccd6353b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.111186] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babd5855-d786-4264-9e17-728c31b5f630 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.142321] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d2a4cc-2416-4375-9615-780736ca146a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.151021] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6798c7-c10f-45e3-892e-2eb2148a2646 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.165518] env[62600]: DEBUG nova.compute.provider_tree [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.192328] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Creating linked-clone VM from snapshot {{(pid=62600) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 953.192950] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d2b2043d-ee8b-4d9e-83ae-0186a7e99785 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.203824] env[62600]: DEBUG oslo_vmware.api [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 953.203824] env[62600]: value = "task-1222804" [ 953.203824] env[62600]: _type = "Task" [ 953.203824] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.215445] env[62600]: DEBUG oslo_vmware.api [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222804, 'name': CloneVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.221043] env[62600]: DEBUG oslo_vmware.api [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222802, 'name': PowerOnVM_Task, 'duration_secs': 0.73482} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.221337] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 953.221562] env[62600]: DEBUG nova.compute.manager [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 953.222487] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c8c6e50-c8f0-40a3-ac39-aeec60f055b0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.517325] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ecff50-a4c4-da54-5e5c-b8a63e33c5d4, 'name': SearchDatastore_Task, 'duration_secs': 0.018239} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.517692] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.517903] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 953.518616] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.518616] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.518616] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 953.518760] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1d3fb2e3-e5f5-4e3f-983e-2bacc9416c8e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.529835] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 953.530048] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 953.530798] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7ace092-7511-43af-a6bb-f59f4076f042 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.539023] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 953.539023] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520cf823-2d29-2168-65db-f023a0a34d70" [ 953.539023] env[62600]: _type = "Task" [ 953.539023] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.554216] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222803, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076421} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.554216] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520cf823-2d29-2168-65db-f023a0a34d70, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.554216] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 953.554928] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed2ed9d-7aac-4170-8e73-0d159060b686 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.584343] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 784e6d19-36ef-4c01-9e9f-5a083ed08608/784e6d19-36ef-4c01-9e9f-5a083ed08608.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 953.584657] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f32f48b5-f127-4d9b-8715-8808d6687a97 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.608440] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 953.608440] env[62600]: value = "task-1222805" [ 953.608440] env[62600]: _type = "Task" [ 953.608440] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.617641] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222805, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.669614] env[62600]: DEBUG nova.scheduler.client.report [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 953.718055] env[62600]: DEBUG oslo_vmware.api [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222804, 'name': CloneVM_Task} progress is 94%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.744350] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.833929] env[62600]: DEBUG nova.compute.manager [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 953.860549] env[62600]: DEBUG nova.virt.hardware [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 953.860815] env[62600]: DEBUG nova.virt.hardware [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 953.860982] env[62600]: DEBUG nova.virt.hardware [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 953.861285] env[62600]: DEBUG nova.virt.hardware [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 953.861379] env[62600]: DEBUG nova.virt.hardware [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 953.861548] env[62600]: DEBUG nova.virt.hardware [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 953.861768] env[62600]: DEBUG nova.virt.hardware [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 953.861933] env[62600]: DEBUG nova.virt.hardware [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 953.862140] env[62600]: DEBUG nova.virt.hardware [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 953.862281] env[62600]: DEBUG nova.virt.hardware [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 953.862464] env[62600]: DEBUG nova.virt.hardware [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 953.863392] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b8dc06c-42f3-4453-8233-f096bfcfd90c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.873662] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bab8bb72-4ed1-43af-a014-dc16dfa9130c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.050242] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520cf823-2d29-2168-65db-f023a0a34d70, 'name': SearchDatastore_Task, 'duration_secs': 0.017842} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.050895] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96d0207b-0691-41f1-ac0d-59b60b10ec0d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.058522] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 954.058522] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5203250d-5177-ec00-28e7-e155da6de29e" [ 954.058522] env[62600]: _type = "Task" [ 954.058522] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.066323] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5203250d-5177-ec00-28e7-e155da6de29e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.118911] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222805, 'name': ReconfigVM_Task, 'duration_secs': 0.450621} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.119364] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 784e6d19-36ef-4c01-9e9f-5a083ed08608/784e6d19-36ef-4c01-9e9f-5a083ed08608.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 954.120051] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-81ded29e-9314-44ef-8c11-7c6cb2c25e21 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.127384] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 954.127384] env[62600]: value = "task-1222806" [ 954.127384] env[62600]: _type = "Task" [ 954.127384] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.137614] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222806, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.174925] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.175509] env[62600]: DEBUG nova.compute.manager [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 954.178423] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.377s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.179970] env[62600]: INFO nova.compute.claims [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 954.216152] env[62600]: DEBUG oslo_vmware.api [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222804, 'name': CloneVM_Task} progress is 94%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.330269] env[62600]: DEBUG nova.compute.manager [req-b577f3f8-f93b-4ee2-b7af-f125683d925f req-20a1c17f-d5ef-4de9-9ed5-7e3a7beb9a4e service nova] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Received event network-vif-plugged-80551113-7315-4224-ac50-2f6b14d0a6b0 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 954.330509] env[62600]: DEBUG oslo_concurrency.lockutils [req-b577f3f8-f93b-4ee2-b7af-f125683d925f req-20a1c17f-d5ef-4de9-9ed5-7e3a7beb9a4e service nova] Acquiring lock "234dc7f8-e2ef-4f20-b25b-dcd61ce20c57-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.330808] env[62600]: DEBUG oslo_concurrency.lockutils [req-b577f3f8-f93b-4ee2-b7af-f125683d925f req-20a1c17f-d5ef-4de9-9ed5-7e3a7beb9a4e service nova] Lock "234dc7f8-e2ef-4f20-b25b-dcd61ce20c57-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.330996] env[62600]: DEBUG oslo_concurrency.lockutils [req-b577f3f8-f93b-4ee2-b7af-f125683d925f req-20a1c17f-d5ef-4de9-9ed5-7e3a7beb9a4e service nova] Lock "234dc7f8-e2ef-4f20-b25b-dcd61ce20c57-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.331321] env[62600]: DEBUG nova.compute.manager [req-b577f3f8-f93b-4ee2-b7af-f125683d925f req-20a1c17f-d5ef-4de9-9ed5-7e3a7beb9a4e service nova] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] No waiting events found dispatching network-vif-plugged-80551113-7315-4224-ac50-2f6b14d0a6b0 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 954.331455] env[62600]: WARNING nova.compute.manager [req-b577f3f8-f93b-4ee2-b7af-f125683d925f req-20a1c17f-d5ef-4de9-9ed5-7e3a7beb9a4e service nova] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Received unexpected event network-vif-plugged-80551113-7315-4224-ac50-2f6b14d0a6b0 for instance with vm_state building and task_state spawning. [ 954.448953] env[62600]: DEBUG nova.network.neutron [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Successfully updated port: 80551113-7315-4224-ac50-2f6b14d0a6b0 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 954.568948] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5203250d-5177-ec00-28e7-e155da6de29e, 'name': SearchDatastore_Task, 'duration_secs': 0.025785} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.569255] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.569491] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 81333628-2616-4557-88d1-0e7164e1b16d/81333628-2616-4557-88d1-0e7164e1b16d.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 954.569768] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-63ff3ac0-0fe5-4bf4-ac96-375e9e88206b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.578888] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 954.578888] env[62600]: value = "task-1222807" [ 954.578888] env[62600]: _type = "Task" [ 954.578888] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.588930] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222807, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.639791] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222806, 'name': Rename_Task, 'duration_secs': 0.225929} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.640140] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 954.640419] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4a1b5036-faef-4514-9f9c-b9acd0689ff4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.649038] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 954.649038] env[62600]: value = "task-1222808" [ 954.649038] env[62600]: _type = "Task" [ 954.649038] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.660326] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222808, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.685572] env[62600]: DEBUG nova.compute.utils [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 954.689888] env[62600]: DEBUG nova.compute.manager [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 954.690116] env[62600]: DEBUG nova.network.neutron [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 954.719081] env[62600]: DEBUG oslo_vmware.api [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222804, 'name': CloneVM_Task} progress is 95%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.797086] env[62600]: DEBUG nova.policy [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '94751040d0fd41a895010c0f2f019131', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef3bb131f3994f7a8673dafff6d353c1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 954.953139] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "refresh_cache-234dc7f8-e2ef-4f20-b25b-dcd61ce20c57" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.955408] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired lock "refresh_cache-234dc7f8-e2ef-4f20-b25b-dcd61ce20c57" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.955408] env[62600]: DEBUG nova.network.neutron [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 955.091424] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222807, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.167747] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222808, 'name': PowerOnVM_Task} progress is 71%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.192109] env[62600]: DEBUG nova.compute.manager [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 955.221660] env[62600]: DEBUG oslo_vmware.api [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222804, 'name': CloneVM_Task, 'duration_secs': 1.762034} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.222457] env[62600]: INFO nova.virt.vmwareapi.vmops [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Created linked-clone VM from snapshot [ 955.223250] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17bef949-55cd-4c90-b2b7-b5e28edae17a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.232946] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Uploading image a4777ff7-01da-48f7-91e7-7530e75e2382 {{(pid=62600) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 955.248628] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Destroying the VM {{(pid=62600) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 955.250531] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-5d754d83-382a-45c6-8582-b81085066284 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.254929] env[62600]: DEBUG nova.network.neutron [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Successfully created port: ead68533-9f71-4ab4-a47d-af3e0c397bd4 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 955.261747] env[62600]: DEBUG oslo_vmware.api [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 955.261747] env[62600]: value = "task-1222809" [ 955.261747] env[62600]: _type = "Task" [ 955.261747] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.274240] env[62600]: DEBUG oslo_vmware.api [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222809, 'name': Destroy_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.391446] env[62600]: DEBUG oslo_concurrency.lockutils [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "f18df958-2885-4d8c-a9ad-140faaed1178" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.391533] env[62600]: DEBUG oslo_concurrency.lockutils [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "f18df958-2885-4d8c-a9ad-140faaed1178" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.391751] env[62600]: DEBUG oslo_concurrency.lockutils [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "f18df958-2885-4d8c-a9ad-140faaed1178-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.391980] env[62600]: DEBUG oslo_concurrency.lockutils [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "f18df958-2885-4d8c-a9ad-140faaed1178-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.392171] env[62600]: DEBUG oslo_concurrency.lockutils [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "f18df958-2885-4d8c-a9ad-140faaed1178-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.395254] env[62600]: INFO nova.compute.manager [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Terminating instance [ 955.398664] env[62600]: DEBUG nova.compute.manager [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 955.398664] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 955.398827] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-114d3ae0-97ef-44ca-8953-bb33183a1b0c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.411221] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 955.411221] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b07454f1-860d-40c6-a146-cef9e4c658af {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.422951] env[62600]: DEBUG oslo_vmware.api [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 955.422951] env[62600]: value = "task-1222810" [ 955.422951] env[62600]: _type = "Task" [ 955.422951] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.432853] env[62600]: DEBUG oslo_vmware.api [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222810, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.491260] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31c4196-d40a-42da-ba3f-93f44490ad74 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.500924] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac62b77f-8fc6-4696-bdd2-4c7000b6515a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.505799] env[62600]: DEBUG nova.network.neutron [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 955.545305] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ad49c87-dff1-4120-bda7-f7581afca14f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.555730] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a42fcdd-0fd1-4c94-b245-93b99fd8bb8b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.572839] env[62600]: DEBUG nova.compute.provider_tree [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.594449] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222807, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.640989} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.594730] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 81333628-2616-4557-88d1-0e7164e1b16d/81333628-2616-4557-88d1-0e7164e1b16d.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 955.594962] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 955.595246] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-536e13ad-0fe8-4962-84b8-e1b7e1057825 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.603032] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 955.603032] env[62600]: value = "task-1222811" [ 955.603032] env[62600]: _type = "Task" [ 955.603032] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.614103] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222811, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.676797] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222808, 'name': PowerOnVM_Task, 'duration_secs': 0.917066} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.677168] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 955.677396] env[62600]: INFO nova.compute.manager [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Took 8.29 seconds to spawn the instance on the hypervisor. [ 955.677598] env[62600]: DEBUG nova.compute.manager [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 955.678550] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d66fecb-b45f-44ce-bbaf-0e6271f63fbb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.709114] env[62600]: DEBUG nova.network.neutron [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Updating instance_info_cache with network_info: [{"id": "80551113-7315-4224-ac50-2f6b14d0a6b0", "address": "fa:16:3e:9f:8c:b6", "network": {"id": "5ef67e8a-5613-4cdb-9f35-0a18720657ff", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1644754500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bab74e3ae78248909dbcd483abbb8da2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80551113-73", "ovs_interfaceid": "80551113-7315-4224-ac50-2f6b14d0a6b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.773811] env[62600]: DEBUG oslo_vmware.api [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222809, 'name': Destroy_Task} progress is 33%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.934054] env[62600]: DEBUG oslo_vmware.api [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222810, 'name': PowerOffVM_Task, 'duration_secs': 0.310144} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.934361] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.934571] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.934836] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9a018f4d-6240-41a0-a391-93ef14f4047a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.003573] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "248a5371-6ff7-4da7-ae97-d638bea0123a" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.003573] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "248a5371-6ff7-4da7-ae97-d638bea0123a" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.076051] env[62600]: DEBUG nova.scheduler.client.report [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.113602] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222811, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.1326} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.113901] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 956.114762] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e21eaaff-6592-4930-8f34-a855c7c38e96 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.132947] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "02c6dac9-535b-436b-bbee-c8031cb5d7f5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.133256] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "02c6dac9-535b-436b-bbee-c8031cb5d7f5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.143223] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 81333628-2616-4557-88d1-0e7164e1b16d/81333628-2616-4557-88d1-0e7164e1b16d.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 956.144075] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-43bf35ec-2f45-451c-bbf1-5f5a6f37b7e0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.162487] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 956.162487] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 956.162487] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleting the datastore file [datastore2] f18df958-2885-4d8c-a9ad-140faaed1178 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.163444] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-32e91681-ce0b-4e9a-8573-92b8fe781075 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.170345] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 956.170345] env[62600]: value = "task-1222813" [ 956.170345] env[62600]: _type = "Task" [ 956.170345] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.171985] env[62600]: DEBUG oslo_vmware.api [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 956.171985] env[62600]: value = "task-1222814" [ 956.171985] env[62600]: _type = "Task" [ 956.171985] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.188552] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222813, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.189143] env[62600]: DEBUG oslo_vmware.api [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222814, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.199215] env[62600]: INFO nova.compute.manager [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Took 28.11 seconds to build instance. [ 956.209559] env[62600]: DEBUG nova.compute.manager [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 956.211979] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Releasing lock "refresh_cache-234dc7f8-e2ef-4f20-b25b-dcd61ce20c57" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.212361] env[62600]: DEBUG nova.compute.manager [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Instance network_info: |[{"id": "80551113-7315-4224-ac50-2f6b14d0a6b0", "address": "fa:16:3e:9f:8c:b6", "network": {"id": "5ef67e8a-5613-4cdb-9f35-0a18720657ff", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1644754500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bab74e3ae78248909dbcd483abbb8da2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80551113-73", "ovs_interfaceid": "80551113-7315-4224-ac50-2f6b14d0a6b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 956.212684] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:8c:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3734b156-0f7d-4721-b23c-d000412ec2eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '80551113-7315-4224-ac50-2f6b14d0a6b0', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 956.222102] env[62600]: DEBUG oslo.service.loopingcall [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.223455] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 956.223904] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6a548fef-7329-43be-912e-525137c9dffa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.246182] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 956.246182] env[62600]: value = "task-1222815" [ 956.246182] env[62600]: _type = "Task" [ 956.246182] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.251702] env[62600]: DEBUG nova.virt.hardware [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 956.251964] env[62600]: DEBUG nova.virt.hardware [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 956.252146] env[62600]: DEBUG nova.virt.hardware [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 956.252345] env[62600]: DEBUG nova.virt.hardware [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 956.252559] env[62600]: DEBUG nova.virt.hardware [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 956.255679] env[62600]: DEBUG nova.virt.hardware [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 956.255679] env[62600]: DEBUG nova.virt.hardware [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 956.255679] env[62600]: DEBUG nova.virt.hardware [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 956.255679] env[62600]: DEBUG nova.virt.hardware [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 956.255679] env[62600]: DEBUG nova.virt.hardware [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 956.255679] env[62600]: DEBUG nova.virt.hardware [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 956.255679] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc963948-0536-4e12-9547-4dbc443b27cc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.262669] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222815, 'name': CreateVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.269011] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-629a7c81-b53b-48d6-a5d5-b4e4585737de {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.279834] env[62600]: DEBUG oslo_vmware.api [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222809, 'name': Destroy_Task, 'duration_secs': 0.655151} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.287991] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Destroyed the VM [ 956.288271] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Deleting Snapshot of the VM instance {{(pid=62600) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 956.289210] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-2a04ef8c-c9a5-449c-947e-66946ca77f81 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.295901] env[62600]: DEBUG oslo_vmware.api [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 956.295901] env[62600]: value = "task-1222816" [ 956.295901] env[62600]: _type = "Task" [ 956.295901] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.304284] env[62600]: DEBUG oslo_vmware.api [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222816, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.506240] env[62600]: DEBUG nova.compute.utils [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 956.509921] env[62600]: DEBUG nova.compute.manager [req-33dad9a3-f419-46c4-ad3b-ec1f6bc65443 req-fa83de52-7b18-4973-ba23-b1b65d7ecc30 service nova] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Received event network-changed-80551113-7315-4224-ac50-2f6b14d0a6b0 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.510013] env[62600]: DEBUG nova.compute.manager [req-33dad9a3-f419-46c4-ad3b-ec1f6bc65443 req-fa83de52-7b18-4973-ba23-b1b65d7ecc30 service nova] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Refreshing instance network info cache due to event network-changed-80551113-7315-4224-ac50-2f6b14d0a6b0. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 956.510222] env[62600]: DEBUG oslo_concurrency.lockutils [req-33dad9a3-f419-46c4-ad3b-ec1f6bc65443 req-fa83de52-7b18-4973-ba23-b1b65d7ecc30 service nova] Acquiring lock "refresh_cache-234dc7f8-e2ef-4f20-b25b-dcd61ce20c57" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.510376] env[62600]: DEBUG oslo_concurrency.lockutils [req-33dad9a3-f419-46c4-ad3b-ec1f6bc65443 req-fa83de52-7b18-4973-ba23-b1b65d7ecc30 service nova] Acquired lock "refresh_cache-234dc7f8-e2ef-4f20-b25b-dcd61ce20c57" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.510497] env[62600]: DEBUG nova.network.neutron [req-33dad9a3-f419-46c4-ad3b-ec1f6bc65443 req-fa83de52-7b18-4973-ba23-b1b65d7ecc30 service nova] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Refreshing network info cache for port 80551113-7315-4224-ac50-2f6b14d0a6b0 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 956.582371] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.582941] env[62600]: DEBUG nova.compute.manager [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 956.586397] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.842s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.590133] env[62600]: DEBUG nova.objects.instance [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62600) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 956.644865] env[62600]: DEBUG nova.compute.manager [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 956.686188] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222813, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.689310] env[62600]: DEBUG oslo_vmware.api [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222814, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.33158} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.689703] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.692080] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 956.692080] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 956.692080] env[62600]: INFO nova.compute.manager [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Took 1.29 seconds to destroy the instance on the hypervisor. [ 956.692080] env[62600]: DEBUG oslo.service.loopingcall [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.692080] env[62600]: DEBUG nova.compute.manager [-] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 956.692080] env[62600]: DEBUG nova.network.neutron [-] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 956.700831] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "784e6d19-36ef-4c01-9e9f-5a083ed08608" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.623s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.758962] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222815, 'name': CreateVM_Task} progress is 25%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.810134] env[62600]: DEBUG oslo_vmware.api [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222816, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.015831] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "248a5371-6ff7-4da7-ae97-d638bea0123a" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.091672] env[62600]: DEBUG nova.compute.utils [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 957.097942] env[62600]: DEBUG nova.compute.manager [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 957.098206] env[62600]: DEBUG nova.network.neutron [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 957.177744] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.183705] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222813, 'name': ReconfigVM_Task, 'duration_secs': 0.520785} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.184019] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 81333628-2616-4557-88d1-0e7164e1b16d/81333628-2616-4557-88d1-0e7164e1b16d.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.185349] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-12064508-34f0-4c23-92b2-6303377ae14b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.192393] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 957.192393] env[62600]: value = "task-1222817" [ 957.192393] env[62600]: _type = "Task" [ 957.192393] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.207671] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222817, 'name': Rename_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.237525] env[62600]: DEBUG nova.policy [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3e59998217a4b18a6f3f01142a5e440', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f91091f83ee4a2091507ca994e3d52f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 957.255973] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222815, 'name': CreateVM_Task, 'duration_secs': 0.830967} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.256655] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 957.257406] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.257584] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.257916] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 957.258202] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-467e39c8-63df-40ca-a7c3-04298ffc28b6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.263117] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 957.263117] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52879da2-0219-03f3-daef-336b1c5d16cf" [ 957.263117] env[62600]: _type = "Task" [ 957.263117] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.272424] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52879da2-0219-03f3-daef-336b1c5d16cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.276538] env[62600]: DEBUG nova.network.neutron [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Successfully updated port: ead68533-9f71-4ab4-a47d-af3e0c397bd4 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 957.310980] env[62600]: DEBUG oslo_vmware.api [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222816, 'name': RemoveSnapshot_Task, 'duration_secs': 0.748268} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.311798] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Deleted Snapshot of the VM instance {{(pid=62600) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 957.576360] env[62600]: DEBUG nova.network.neutron [req-33dad9a3-f419-46c4-ad3b-ec1f6bc65443 req-fa83de52-7b18-4973-ba23-b1b65d7ecc30 service nova] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Updated VIF entry in instance network info cache for port 80551113-7315-4224-ac50-2f6b14d0a6b0. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 957.576998] env[62600]: DEBUG nova.network.neutron [req-33dad9a3-f419-46c4-ad3b-ec1f6bc65443 req-fa83de52-7b18-4973-ba23-b1b65d7ecc30 service nova] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Updating instance_info_cache with network_info: [{"id": "80551113-7315-4224-ac50-2f6b14d0a6b0", "address": "fa:16:3e:9f:8c:b6", "network": {"id": "5ef67e8a-5613-4cdb-9f35-0a18720657ff", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1644754500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bab74e3ae78248909dbcd483abbb8da2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80551113-73", "ovs_interfaceid": "80551113-7315-4224-ac50-2f6b14d0a6b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.598827] env[62600]: DEBUG nova.compute.manager [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 957.604770] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e6305f1f-76d3-4938-9d0f-e90c3416dfa3 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.605814] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.428s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.607870] env[62600]: INFO nova.compute.claims [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 957.662192] env[62600]: DEBUG oslo_vmware.rw_handles [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f06b44-4035-1d45-71c8-ffa1a25bc384/disk-0.vmdk. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 957.663440] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90cb49b9-b0fa-454e-8a74-683a1b6945de {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.672330] env[62600]: DEBUG oslo_vmware.rw_handles [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f06b44-4035-1d45-71c8-ffa1a25bc384/disk-0.vmdk is in state: ready. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 957.672568] env[62600]: ERROR oslo_vmware.rw_handles [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f06b44-4035-1d45-71c8-ffa1a25bc384/disk-0.vmdk due to incomplete transfer. [ 957.672899] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b5fc6586-a060-478d-8103-3f8109ae2cdf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.684439] env[62600]: DEBUG oslo_vmware.rw_handles [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52f06b44-4035-1d45-71c8-ffa1a25bc384/disk-0.vmdk. {{(pid=62600) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 957.684680] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Uploaded image bc24c6fc-64cd-481c-956d-7fa0f16f714f to the Glance image server {{(pid=62600) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 957.687954] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Destroying the VM {{(pid=62600) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 957.688522] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-7989dac8-8b27-4e83-84f1-01db31d0b54f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.697806] env[62600]: DEBUG oslo_vmware.api [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for the task: (returnval){ [ 957.697806] env[62600]: value = "task-1222818" [ 957.697806] env[62600]: _type = "Task" [ 957.697806] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.704226] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222817, 'name': Rename_Task, 'duration_secs': 0.225138} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.705181] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 957.705445] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8ddffa7a-add2-4d35-855d-af9346dd7c0a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.712906] env[62600]: DEBUG oslo_vmware.api [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222818, 'name': Destroy_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.716840] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 957.716840] env[62600]: value = "task-1222819" [ 957.716840] env[62600]: _type = "Task" [ 957.716840] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.723996] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222819, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.774301] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52879da2-0219-03f3-daef-336b1c5d16cf, 'name': SearchDatastore_Task, 'duration_secs': 0.015617} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.775211] env[62600]: DEBUG nova.network.neutron [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Successfully created port: a3a87739-c27f-4a13-a4f3-90ac701486e8 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 957.777343] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.777587] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 957.777865] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.778040] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.778236] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 957.778517] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0c9ead13-9957-4e28-8b48-2d47932c2693 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.783435] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "refresh_cache-39deb498-6bf0-4f3b-932b-8068fc48271e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.783536] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquired lock "refresh_cache-39deb498-6bf0-4f3b-932b-8068fc48271e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.783642] env[62600]: DEBUG nova.network.neutron [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 957.791292] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 957.791292] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 957.791292] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7db7ec91-e11e-4ad5-ae00-fce9e38b4ff6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.796586] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 957.796586] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5278250f-9285-05ae-fc85-7e1a6532f577" [ 957.796586] env[62600]: _type = "Task" [ 957.796586] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.805636] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5278250f-9285-05ae-fc85-7e1a6532f577, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.820462] env[62600]: WARNING nova.compute.manager [None req-d7b7ddf8-dce5-4a55-a9d7-1d0df9da2372 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Image not found during snapshot: nova.exception.ImageNotFound: Image a4777ff7-01da-48f7-91e7-7530e75e2382 could not be found. [ 957.831463] env[62600]: DEBUG nova.network.neutron [-] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.084104] env[62600]: DEBUG oslo_concurrency.lockutils [req-33dad9a3-f419-46c4-ad3b-ec1f6bc65443 req-fa83de52-7b18-4973-ba23-b1b65d7ecc30 service nova] Releasing lock "refresh_cache-234dc7f8-e2ef-4f20-b25b-dcd61ce20c57" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.094380] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "248a5371-6ff7-4da7-ae97-d638bea0123a" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.094690] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "248a5371-6ff7-4da7-ae97-d638bea0123a" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.095067] env[62600]: INFO nova.compute.manager [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Attaching volume e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf to /dev/sdb [ 958.147378] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b019bfe6-312c-4004-b49f-9a45c51e329e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.155959] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02ab46a0-c5a6-4ad0-8e3b-885aff22d32a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.172106] env[62600]: DEBUG nova.virt.block_device [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Updating existing volume attachment record: f8ab2ccb-1ed5-45a7-8b24-70d2d8f95444 {{(pid=62600) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 958.210022] env[62600]: DEBUG oslo_vmware.api [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222818, 'name': Destroy_Task, 'duration_secs': 0.35212} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.210319] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Destroyed the VM [ 958.210778] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Deleting Snapshot of the VM instance {{(pid=62600) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 958.211104] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-42dd810c-d724-4d5b-83a1-8a21ecc00815 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.219096] env[62600]: DEBUG oslo_vmware.api [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for the task: (returnval){ [ 958.219096] env[62600]: value = "task-1222820" [ 958.219096] env[62600]: _type = "Task" [ 958.219096] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.231798] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222819, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.235173] env[62600]: DEBUG oslo_vmware.api [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222820, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.307764] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5278250f-9285-05ae-fc85-7e1a6532f577, 'name': SearchDatastore_Task, 'duration_secs': 0.020788} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.308727] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b05a44c5-d191-4157-8b21-327b191466e6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.314949] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 958.314949] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5231eed8-3f36-5a63-4e1e-13d9a802f7e8" [ 958.314949] env[62600]: _type = "Task" [ 958.314949] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.325943] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5231eed8-3f36-5a63-4e1e-13d9a802f7e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.328624] env[62600]: DEBUG nova.network.neutron [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 958.339954] env[62600]: INFO nova.compute.manager [-] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Took 1.65 seconds to deallocate network for instance. [ 958.612279] env[62600]: DEBUG nova.network.neutron [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Updating instance_info_cache with network_info: [{"id": "ead68533-9f71-4ab4-a47d-af3e0c397bd4", "address": "fa:16:3e:f6:16:4c", "network": {"id": "e7547ac4-24a6-4778-a17a-40e79e7a0d0c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-320671681-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3bb131f3994f7a8673dafff6d353c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapead68533-9f", "ovs_interfaceid": "ead68533-9f71-4ab4-a47d-af3e0c397bd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.635837] env[62600]: DEBUG nova.compute.manager [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 958.656695] env[62600]: DEBUG nova.compute.manager [req-cfb2a11c-90d8-4bcd-9062-3e77b6b59e47 req-900df6b1-88b7-4d85-93f5-ddb8adca379d service nova] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Received event network-vif-plugged-ead68533-9f71-4ab4-a47d-af3e0c397bd4 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.656849] env[62600]: DEBUG oslo_concurrency.lockutils [req-cfb2a11c-90d8-4bcd-9062-3e77b6b59e47 req-900df6b1-88b7-4d85-93f5-ddb8adca379d service nova] Acquiring lock "39deb498-6bf0-4f3b-932b-8068fc48271e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.657426] env[62600]: DEBUG oslo_concurrency.lockutils [req-cfb2a11c-90d8-4bcd-9062-3e77b6b59e47 req-900df6b1-88b7-4d85-93f5-ddb8adca379d service nova] Lock "39deb498-6bf0-4f3b-932b-8068fc48271e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.657426] env[62600]: DEBUG oslo_concurrency.lockutils [req-cfb2a11c-90d8-4bcd-9062-3e77b6b59e47 req-900df6b1-88b7-4d85-93f5-ddb8adca379d service nova] Lock "39deb498-6bf0-4f3b-932b-8068fc48271e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.658090] env[62600]: DEBUG nova.compute.manager [req-cfb2a11c-90d8-4bcd-9062-3e77b6b59e47 req-900df6b1-88b7-4d85-93f5-ddb8adca379d service nova] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] No waiting events found dispatching network-vif-plugged-ead68533-9f71-4ab4-a47d-af3e0c397bd4 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 958.658339] env[62600]: WARNING nova.compute.manager [req-cfb2a11c-90d8-4bcd-9062-3e77b6b59e47 req-900df6b1-88b7-4d85-93f5-ddb8adca379d service nova] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Received unexpected event network-vif-plugged-ead68533-9f71-4ab4-a47d-af3e0c397bd4 for instance with vm_state building and task_state spawning. [ 958.659061] env[62600]: DEBUG nova.compute.manager [req-cfb2a11c-90d8-4bcd-9062-3e77b6b59e47 req-900df6b1-88b7-4d85-93f5-ddb8adca379d service nova] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Received event network-changed-ead68533-9f71-4ab4-a47d-af3e0c397bd4 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.659061] env[62600]: DEBUG nova.compute.manager [req-cfb2a11c-90d8-4bcd-9062-3e77b6b59e47 req-900df6b1-88b7-4d85-93f5-ddb8adca379d service nova] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Refreshing instance network info cache due to event network-changed-ead68533-9f71-4ab4-a47d-af3e0c397bd4. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 958.659375] env[62600]: DEBUG oslo_concurrency.lockutils [req-cfb2a11c-90d8-4bcd-9062-3e77b6b59e47 req-900df6b1-88b7-4d85-93f5-ddb8adca379d service nova] Acquiring lock "refresh_cache-39deb498-6bf0-4f3b-932b-8068fc48271e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.662167] env[62600]: DEBUG nova.virt.hardware [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 958.662432] env[62600]: DEBUG nova.virt.hardware [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 958.662632] env[62600]: DEBUG nova.virt.hardware [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 958.662851] env[62600]: DEBUG nova.virt.hardware [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 958.663038] env[62600]: DEBUG nova.virt.hardware [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 958.663205] env[62600]: DEBUG nova.virt.hardware [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 958.663481] env[62600]: DEBUG nova.virt.hardware [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 958.663659] env[62600]: DEBUG nova.virt.hardware [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 958.663837] env[62600]: DEBUG nova.virt.hardware [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 958.664019] env[62600]: DEBUG nova.virt.hardware [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 958.664928] env[62600]: DEBUG nova.virt.hardware [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 958.665831] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-472dfb2f-2b23-4f32-8763-c426b2329514 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.679054] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5a6bcfd-2c6a-46a1-9697-a53399e348df {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.734544] env[62600]: DEBUG oslo_vmware.api [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222819, 'name': PowerOnVM_Task, 'duration_secs': 0.668149} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.735625] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 958.735958] env[62600]: INFO nova.compute.manager [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Took 9.05 seconds to spawn the instance on the hypervisor. [ 958.736276] env[62600]: DEBUG nova.compute.manager [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 958.745285] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5c3feeb-2b30-427f-ac50-d10f447822e6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.754027] env[62600]: DEBUG oslo_vmware.api [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222820, 'name': RemoveSnapshot_Task} progress is 70%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.826589] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5231eed8-3f36-5a63-4e1e-13d9a802f7e8, 'name': SearchDatastore_Task, 'duration_secs': 0.022853} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.829248] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.830383] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57/234dc7f8-e2ef-4f20-b25b-dcd61ce20c57.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 958.832689] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9c81c54e-26c4-4e53-bbbb-1e547c3c7fdb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.839338] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 958.839338] env[62600]: value = "task-1222824" [ 958.839338] env[62600]: _type = "Task" [ 958.839338] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.850112] env[62600]: DEBUG oslo_concurrency.lockutils [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.851015] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222824, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.960022] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8b02e3b-c9ff-4094-a624-3e514ec4bee3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.968027] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffbfd630-0aa2-427b-ae13-6462836cdbea {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.000623] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde53c73-4c64-482b-a3a4-ac8adb239813 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.008950] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a54b9d0b-0e26-4b71-ae30-2be833ab0062 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.022961] env[62600]: DEBUG nova.compute.provider_tree [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.062874] env[62600]: DEBUG oslo_concurrency.lockutils [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "84dc284d-c7f4-4fc2-b539-8d4820429b02" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.063695] env[62600]: DEBUG oslo_concurrency.lockutils [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "84dc284d-c7f4-4fc2-b539-8d4820429b02" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.063695] env[62600]: DEBUG oslo_concurrency.lockutils [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "84dc284d-c7f4-4fc2-b539-8d4820429b02-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.063695] env[62600]: DEBUG oslo_concurrency.lockutils [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "84dc284d-c7f4-4fc2-b539-8d4820429b02-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.063959] env[62600]: DEBUG oslo_concurrency.lockutils [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "84dc284d-c7f4-4fc2-b539-8d4820429b02-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.066929] env[62600]: INFO nova.compute.manager [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Terminating instance [ 959.067902] env[62600]: DEBUG nova.compute.manager [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 959.068106] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 959.068917] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98acda23-7ce9-413d-afde-38181735c77a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.077215] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 959.077462] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eb1585f8-1819-4a8c-bdbe-35990668573c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.088940] env[62600]: DEBUG oslo_vmware.api [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 959.088940] env[62600]: value = "task-1222825" [ 959.088940] env[62600]: _type = "Task" [ 959.088940] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.096867] env[62600]: DEBUG oslo_vmware.api [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222825, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.115086] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Releasing lock "refresh_cache-39deb498-6bf0-4f3b-932b-8068fc48271e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.115509] env[62600]: DEBUG nova.compute.manager [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Instance network_info: |[{"id": "ead68533-9f71-4ab4-a47d-af3e0c397bd4", "address": "fa:16:3e:f6:16:4c", "network": {"id": "e7547ac4-24a6-4778-a17a-40e79e7a0d0c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-320671681-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3bb131f3994f7a8673dafff6d353c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapead68533-9f", "ovs_interfaceid": "ead68533-9f71-4ab4-a47d-af3e0c397bd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 959.115869] env[62600]: DEBUG oslo_concurrency.lockutils [req-cfb2a11c-90d8-4bcd-9062-3e77b6b59e47 req-900df6b1-88b7-4d85-93f5-ddb8adca379d service nova] Acquired lock "refresh_cache-39deb498-6bf0-4f3b-932b-8068fc48271e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.116080] env[62600]: DEBUG nova.network.neutron [req-cfb2a11c-90d8-4bcd-9062-3e77b6b59e47 req-900df6b1-88b7-4d85-93f5-ddb8adca379d service nova] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Refreshing network info cache for port ead68533-9f71-4ab4-a47d-af3e0c397bd4 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 959.117528] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:16:4c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ead68533-9f71-4ab4-a47d-af3e0c397bd4', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 959.125668] env[62600]: DEBUG oslo.service.loopingcall [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.126264] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 959.126514] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3c5b5c69-2105-46aa-8510-d91841d297b5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.150754] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 959.150754] env[62600]: value = "task-1222826" [ 959.150754] env[62600]: _type = "Task" [ 959.150754] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.162256] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222826, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.239085] env[62600]: DEBUG oslo_vmware.api [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222820, 'name': RemoveSnapshot_Task, 'duration_secs': 0.928182} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.239085] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Deleted Snapshot of the VM instance {{(pid=62600) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 959.239085] env[62600]: INFO nova.compute.manager [None req-3e845c7a-cdaf-4699-ac23-c56cd424c1ae tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Took 14.53 seconds to snapshot the instance on the hypervisor. [ 959.267867] env[62600]: INFO nova.compute.manager [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Took 31.12 seconds to build instance. [ 959.354596] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222824, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.530210] env[62600]: DEBUG nova.scheduler.client.report [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 959.601689] env[62600]: DEBUG oslo_vmware.api [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222825, 'name': PowerOffVM_Task, 'duration_secs': 0.211757} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.601965] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 959.605659] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 959.605659] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-342c01c1-9e23-4e2c-812e-1382f0d381ce {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.613102] env[62600]: DEBUG nova.compute.manager [req-e30710ba-48fa-448f-84bb-d2ba4a11d6a3 req-5c0a1f82-72f3-4ca9-b733-f97023358c08 service nova] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Received event network-vif-plugged-a3a87739-c27f-4a13-a4f3-90ac701486e8 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 959.613102] env[62600]: DEBUG oslo_concurrency.lockutils [req-e30710ba-48fa-448f-84bb-d2ba4a11d6a3 req-5c0a1f82-72f3-4ca9-b733-f97023358c08 service nova] Acquiring lock "4d12c133-84d7-4ce9-83db-e4b70b8947f3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.613102] env[62600]: DEBUG oslo_concurrency.lockutils [req-e30710ba-48fa-448f-84bb-d2ba4a11d6a3 req-5c0a1f82-72f3-4ca9-b733-f97023358c08 service nova] Lock "4d12c133-84d7-4ce9-83db-e4b70b8947f3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.613102] env[62600]: DEBUG oslo_concurrency.lockutils [req-e30710ba-48fa-448f-84bb-d2ba4a11d6a3 req-5c0a1f82-72f3-4ca9-b733-f97023358c08 service nova] Lock "4d12c133-84d7-4ce9-83db-e4b70b8947f3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.613102] env[62600]: DEBUG nova.compute.manager [req-e30710ba-48fa-448f-84bb-d2ba4a11d6a3 req-5c0a1f82-72f3-4ca9-b733-f97023358c08 service nova] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] No waiting events found dispatching network-vif-plugged-a3a87739-c27f-4a13-a4f3-90ac701486e8 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 959.613102] env[62600]: WARNING nova.compute.manager [req-e30710ba-48fa-448f-84bb-d2ba4a11d6a3 req-5c0a1f82-72f3-4ca9-b733-f97023358c08 service nova] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Received unexpected event network-vif-plugged-a3a87739-c27f-4a13-a4f3-90ac701486e8 for instance with vm_state building and task_state spawning. [ 959.663310] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222826, 'name': CreateVM_Task, 'duration_secs': 0.367377} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.663979] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 959.664376] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.664666] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.665158] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 959.665482] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a275d84-8db9-4911-ad9c-6e9714fecc73 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.671375] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 959.671560] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 959.672695] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Deleting the datastore file [datastore2] 84dc284d-c7f4-4fc2-b539-8d4820429b02 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 959.673253] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4de69a6e-9042-4fb6-9a7b-afe1acbc6efc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.675742] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 959.675742] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]522d9e1c-9e55-6f1c-b8b9-7824f9a8e453" [ 959.675742] env[62600]: _type = "Task" [ 959.675742] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.684077] env[62600]: DEBUG oslo_vmware.api [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for the task: (returnval){ [ 959.684077] env[62600]: value = "task-1222828" [ 959.684077] env[62600]: _type = "Task" [ 959.684077] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.692933] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]522d9e1c-9e55-6f1c-b8b9-7824f9a8e453, 'name': SearchDatastore_Task, 'duration_secs': 0.014455} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.693861] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.694239] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 959.694732] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.694902] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.695221] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 959.698992] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ffb67362-321d-4a7b-8750-a5f18d928a09 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.701106] env[62600]: DEBUG oslo_vmware.api [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222828, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.711076] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 959.711076] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 959.711076] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c0d3425-ca93-4aae-b084-4649ef8a2679 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.721209] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 959.721209] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d1039e-44eb-8567-1146-b469ab872960" [ 959.721209] env[62600]: _type = "Task" [ 959.721209] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.736496] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d1039e-44eb-8567-1146-b469ab872960, 'name': SearchDatastore_Task, 'duration_secs': 0.015048} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.737103] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ba8bddf-dfe1-464e-9274-4aef47262bfc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.745082] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 959.745082] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528be6c3-543a-deed-1d8c-956ff28d4b9b" [ 959.745082] env[62600]: _type = "Task" [ 959.745082] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.756969] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528be6c3-543a-deed-1d8c-956ff28d4b9b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.770800] env[62600]: DEBUG oslo_concurrency.lockutils [None req-721d1739-89e9-45c8-8715-673bdcbc4bed tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "81333628-2616-4557-88d1-0e7164e1b16d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.634s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.852369] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222824, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.631004} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.852655] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57/234dc7f8-e2ef-4f20-b25b-dcd61ce20c57.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 959.852957] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 959.853154] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7fe7a785-8d93-45c4-bc2c-5ea49adc1e63 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.857165] env[62600]: DEBUG nova.network.neutron [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Successfully updated port: a3a87739-c27f-4a13-a4f3-90ac701486e8 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 959.859728] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 959.859728] env[62600]: value = "task-1222829" [ 959.859728] env[62600]: _type = "Task" [ 959.859728] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.868127] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222829, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.036519] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.431s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.037081] env[62600]: DEBUG nova.compute.manager [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 960.039972] env[62600]: DEBUG oslo_concurrency.lockutils [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.190s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.040141] env[62600]: DEBUG nova.objects.instance [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lazy-loading 'resources' on Instance uuid f18df958-2885-4d8c-a9ad-140faaed1178 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.090846] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "784e6d19-36ef-4c01-9e9f-5a083ed08608" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.091134] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "784e6d19-36ef-4c01-9e9f-5a083ed08608" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.091348] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "784e6d19-36ef-4c01-9e9f-5a083ed08608-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.091540] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "784e6d19-36ef-4c01-9e9f-5a083ed08608-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.091761] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "784e6d19-36ef-4c01-9e9f-5a083ed08608-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.094110] env[62600]: INFO nova.compute.manager [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Terminating instance [ 960.095972] env[62600]: DEBUG nova.compute.manager [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 960.096209] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 960.097064] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e86bdcf0-13bd-4239-976f-4b35efafebd8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.104939] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 960.105190] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a10fef94-fbd1-4063-8ffb-cedfe313f866 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.111042] env[62600]: DEBUG oslo_vmware.api [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 960.111042] env[62600]: value = "task-1222830" [ 960.111042] env[62600]: _type = "Task" [ 960.111042] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.119435] env[62600]: DEBUG oslo_vmware.api [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222830, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.141600] env[62600]: DEBUG nova.network.neutron [req-cfb2a11c-90d8-4bcd-9062-3e77b6b59e47 req-900df6b1-88b7-4d85-93f5-ddb8adca379d service nova] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Updated VIF entry in instance network info cache for port ead68533-9f71-4ab4-a47d-af3e0c397bd4. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 960.141984] env[62600]: DEBUG nova.network.neutron [req-cfb2a11c-90d8-4bcd-9062-3e77b6b59e47 req-900df6b1-88b7-4d85-93f5-ddb8adca379d service nova] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Updating instance_info_cache with network_info: [{"id": "ead68533-9f71-4ab4-a47d-af3e0c397bd4", "address": "fa:16:3e:f6:16:4c", "network": {"id": "e7547ac4-24a6-4778-a17a-40e79e7a0d0c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-320671681-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3bb131f3994f7a8673dafff6d353c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapead68533-9f", "ovs_interfaceid": "ead68533-9f71-4ab4-a47d-af3e0c397bd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.195091] env[62600]: DEBUG oslo_vmware.api [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Task: {'id': task-1222828, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.47707} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.195611] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 960.195812] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 960.196080] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 960.196285] env[62600]: INFO nova.compute.manager [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Took 1.13 seconds to destroy the instance on the hypervisor. [ 960.196541] env[62600]: DEBUG oslo.service.loopingcall [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 960.196744] env[62600]: DEBUG nova.compute.manager [-] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 960.196839] env[62600]: DEBUG nova.network.neutron [-] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 960.254785] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528be6c3-543a-deed-1d8c-956ff28d4b9b, 'name': SearchDatastore_Task, 'duration_secs': 0.022053} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.255099] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.255366] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 39deb498-6bf0-4f3b-932b-8068fc48271e/39deb498-6bf0-4f3b-932b-8068fc48271e.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 960.255634] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5aebf1db-1ab4-4419-9552-a52216628af9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.262202] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 960.262202] env[62600]: value = "task-1222831" [ 960.262202] env[62600]: _type = "Task" [ 960.262202] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.270534] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222831, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.360949] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "refresh_cache-4d12c133-84d7-4ce9-83db-e4b70b8947f3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.361120] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "refresh_cache-4d12c133-84d7-4ce9-83db-e4b70b8947f3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.361280] env[62600]: DEBUG nova.network.neutron [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 960.371277] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222829, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082703} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.372233] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 960.372934] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c65114dd-c25e-476e-96bf-3c62e881a800 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.377278] env[62600]: DEBUG oslo_concurrency.lockutils [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "81333628-2616-4557-88d1-0e7164e1b16d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.377540] env[62600]: DEBUG oslo_concurrency.lockutils [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "81333628-2616-4557-88d1-0e7164e1b16d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.377761] env[62600]: DEBUG oslo_concurrency.lockutils [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "81333628-2616-4557-88d1-0e7164e1b16d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.377958] env[62600]: DEBUG oslo_concurrency.lockutils [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "81333628-2616-4557-88d1-0e7164e1b16d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.378165] env[62600]: DEBUG oslo_concurrency.lockutils [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "81333628-2616-4557-88d1-0e7164e1b16d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.398356] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57/234dc7f8-e2ef-4f20-b25b-dcd61ce20c57.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 960.398972] env[62600]: INFO nova.compute.manager [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Terminating instance [ 960.400686] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-070ab051-3082-406c-84d1-a1391ef1f16f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.416853] env[62600]: DEBUG nova.compute.manager [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 960.417100] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 960.418342] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad22249-2c0e-4207-b3b3-7391c980af7a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.426724] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 960.427902] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4afd008d-7b7b-4e1a-a305-15fd3d0bd462 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.429556] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 960.429556] env[62600]: value = "task-1222832" [ 960.429556] env[62600]: _type = "Task" [ 960.429556] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.434249] env[62600]: DEBUG oslo_vmware.api [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 960.434249] env[62600]: value = "task-1222833" [ 960.434249] env[62600]: _type = "Task" [ 960.434249] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.437610] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222832, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.445135] env[62600]: DEBUG oslo_vmware.api [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222833, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.543573] env[62600]: DEBUG nova.compute.utils [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 960.545200] env[62600]: DEBUG nova.compute.manager [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 960.545396] env[62600]: DEBUG nova.network.neutron [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 960.631803] env[62600]: DEBUG oslo_vmware.api [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222830, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.645981] env[62600]: DEBUG oslo_concurrency.lockutils [req-cfb2a11c-90d8-4bcd-9062-3e77b6b59e47 req-900df6b1-88b7-4d85-93f5-ddb8adca379d service nova] Releasing lock "refresh_cache-39deb498-6bf0-4f3b-932b-8068fc48271e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.646349] env[62600]: DEBUG nova.compute.manager [req-cfb2a11c-90d8-4bcd-9062-3e77b6b59e47 req-900df6b1-88b7-4d85-93f5-ddb8adca379d service nova] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Received event network-vif-deleted-874235d5-d1b2-4382-a30a-2e187141de5b {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.697236] env[62600]: DEBUG nova.policy [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bd0d91fa2ac74c38b9e0ebf39c3baa7a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73e664d178f7484a9f4741b4d9450e68', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 960.704434] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Acquiring lock "9ed08410-a4cb-43c4-9271-2c04d9c87eac" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.704712] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Lock "9ed08410-a4cb-43c4-9271-2c04d9c87eac" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.704929] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Acquiring lock "9ed08410-a4cb-43c4-9271-2c04d9c87eac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.705146] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Lock "9ed08410-a4cb-43c4-9271-2c04d9c87eac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.705326] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Lock "9ed08410-a4cb-43c4-9271-2c04d9c87eac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.707772] env[62600]: INFO nova.compute.manager [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Terminating instance [ 960.710683] env[62600]: DEBUG nova.compute.manager [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 960.710790] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 960.711742] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55beefec-03b9-4345-8783-d854f23b885a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.721231] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 960.725135] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d491cd2a-484e-4ca1-9602-fafbbe0e2ce5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.738239] env[62600]: DEBUG oslo_vmware.api [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for the task: (returnval){ [ 960.738239] env[62600]: value = "task-1222835" [ 960.738239] env[62600]: _type = "Task" [ 960.738239] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.753926] env[62600]: DEBUG oslo_vmware.api [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222835, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.774770] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222831, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.912877] env[62600]: DEBUG nova.network.neutron [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 960.931926] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f8f5324-fb2d-4bc3-98da-a612d839c7b6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.951046] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88989112-4a75-43f8-a488-37c021c9b477 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.959695] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222832, 'name': ReconfigVM_Task, 'duration_secs': 0.329445} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.960091] env[62600]: DEBUG oslo_vmware.api [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222833, 'name': PowerOffVM_Task, 'duration_secs': 0.29956} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.960778] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57/234dc7f8-e2ef-4f20-b25b-dcd61ce20c57.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 960.961672] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 960.961924] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 960.962175] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-14ac4f2b-0c13-4c28-9891-09015193efa7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.963931] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-57e27ece-6836-48a9-8359-dfe8c5ac843b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.998954] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31171fa7-2532-44b0-b786-ce8bfefd08b9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.003500] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 961.003500] env[62600]: value = "task-1222836" [ 961.003500] env[62600]: _type = "Task" [ 961.003500] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.010670] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d939e9b8-0d3f-4d52-97f6-fd589c131f98 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.020082] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222836, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.028427] env[62600]: DEBUG nova.compute.provider_tree [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.054543] env[62600]: DEBUG nova.compute.manager [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 961.061214] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 961.061378] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 961.061608] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Deleting the datastore file [datastore2] 81333628-2616-4557-88d1-0e7164e1b16d {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 961.061886] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e3de3ccf-5c8a-4271-a42e-3d7f68a22fc3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.068490] env[62600]: DEBUG oslo_vmware.api [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 961.068490] env[62600]: value = "task-1222838" [ 961.068490] env[62600]: _type = "Task" [ 961.068490] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.076741] env[62600]: DEBUG oslo_vmware.api [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222838, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.117763] env[62600]: DEBUG nova.network.neutron [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Updating instance_info_cache with network_info: [{"id": "a3a87739-c27f-4a13-a4f3-90ac701486e8", "address": "fa:16:3e:f8:77:26", "network": {"id": "1d2052cf-5be1-48e7-887a-18f367b7823f", "bridge": "br-int", "label": "tempest-ServersTestJSON-106099679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f91091f83ee4a2091507ca994e3d52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3a87739-c2", "ovs_interfaceid": "a3a87739-c27f-4a13-a4f3-90ac701486e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.125755] env[62600]: DEBUG oslo_vmware.api [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222830, 'name': PowerOffVM_Task, 'duration_secs': 0.61486} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.126113] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 961.126302] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 961.126703] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a21a1fd1-692a-4115-9460-a90a387e7c15 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.189949] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 961.192133] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 961.192133] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Deleting the datastore file [datastore2] 784e6d19-36ef-4c01-9e9f-5a083ed08608 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 961.192133] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-032e0e0d-dfa7-471e-a2e4-c90642cab561 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.196610] env[62600]: DEBUG oslo_vmware.api [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for the task: (returnval){ [ 961.196610] env[62600]: value = "task-1222840" [ 961.196610] env[62600]: _type = "Task" [ 961.196610] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.210020] env[62600]: DEBUG oslo_vmware.api [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222840, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.247857] env[62600]: DEBUG oslo_vmware.api [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222835, 'name': PowerOffVM_Task, 'duration_secs': 0.276172} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.248163] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 961.248360] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 961.248610] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-28e77fa7-afa3-4c61-b0b4-edf1980331bc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.272736] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222831, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.681957} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.272997] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 39deb498-6bf0-4f3b-932b-8068fc48271e/39deb498-6bf0-4f3b-932b-8068fc48271e.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 961.273260] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 961.274042] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9bd1da9f-cb02-4a87-8f82-ac094581131d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.280374] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 961.280374] env[62600]: value = "task-1222842" [ 961.280374] env[62600]: _type = "Task" [ 961.280374] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.288632] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222842, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.315265] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 961.315509] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 961.315777] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Deleting the datastore file [datastore2] 9ed08410-a4cb-43c4-9271-2c04d9c87eac {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 961.315999] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-759bedce-9aea-468f-a80b-11165caf6bc3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.324143] env[62600]: DEBUG oslo_vmware.api [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for the task: (returnval){ [ 961.324143] env[62600]: value = "task-1222843" [ 961.324143] env[62600]: _type = "Task" [ 961.324143] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.331890] env[62600]: DEBUG oslo_vmware.api [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222843, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.400640] env[62600]: DEBUG nova.network.neutron [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Successfully created port: 3753b6a6-f27d-4e1d-b801-91324a7e4e0e {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 961.515457] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222836, 'name': Rename_Task, 'duration_secs': 0.184014} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.515754] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 961.516010] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a8bf6f7f-d813-420d-888e-10b0b91d9d48 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.522120] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 961.522120] env[62600]: value = "task-1222844" [ 961.522120] env[62600]: _type = "Task" [ 961.522120] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.529991] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222844, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.530832] env[62600]: DEBUG nova.scheduler.client.report [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 961.534483] env[62600]: DEBUG nova.network.neutron [-] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.579744] env[62600]: DEBUG oslo_vmware.api [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222838, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157015} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.580103] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 961.580234] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 961.580406] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 961.580588] env[62600]: INFO nova.compute.manager [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Took 1.16 seconds to destroy the instance on the hypervisor. [ 961.580831] env[62600]: DEBUG oslo.service.loopingcall [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.581039] env[62600]: DEBUG nova.compute.manager [-] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 961.581143] env[62600]: DEBUG nova.network.neutron [-] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 961.621077] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "refresh_cache-4d12c133-84d7-4ce9-83db-e4b70b8947f3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.621633] env[62600]: DEBUG nova.compute.manager [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Instance network_info: |[{"id": "a3a87739-c27f-4a13-a4f3-90ac701486e8", "address": "fa:16:3e:f8:77:26", "network": {"id": "1d2052cf-5be1-48e7-887a-18f367b7823f", "bridge": "br-int", "label": "tempest-ServersTestJSON-106099679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f91091f83ee4a2091507ca994e3d52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3a87739-c2", "ovs_interfaceid": "a3a87739-c27f-4a13-a4f3-90ac701486e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 961.622118] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f8:77:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '028bae2d-fe6c-4207-b4a3-3fab45fbf1d6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a3a87739-c27f-4a13-a4f3-90ac701486e8', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 961.631037] env[62600]: DEBUG oslo.service.loopingcall [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.631295] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 961.631548] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1f1a0b7a-0433-4e13-9c73-ed8a36d8e83f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.653991] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 961.653991] env[62600]: value = "task-1222845" [ 961.653991] env[62600]: _type = "Task" [ 961.653991] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.669454] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222845, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.706132] env[62600]: DEBUG oslo_vmware.api [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Task: {'id': task-1222840, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.194187} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.706413] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 961.706613] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 961.706799] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 961.706981] env[62600]: INFO nova.compute.manager [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Took 1.61 seconds to destroy the instance on the hypervisor. [ 961.707250] env[62600]: DEBUG oslo.service.loopingcall [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.707447] env[62600]: DEBUG nova.compute.manager [-] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 961.707546] env[62600]: DEBUG nova.network.neutron [-] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 961.792018] env[62600]: DEBUG nova.compute.manager [req-1ce6075e-c335-4acf-892d-a83071c8c4de req-7b444df9-0bb1-43cc-b444-b4f7ab0d498f service nova] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Received event network-changed-a3a87739-c27f-4a13-a4f3-90ac701486e8 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 961.792018] env[62600]: DEBUG nova.compute.manager [req-1ce6075e-c335-4acf-892d-a83071c8c4de req-7b444df9-0bb1-43cc-b444-b4f7ab0d498f service nova] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Refreshing instance network info cache due to event network-changed-a3a87739-c27f-4a13-a4f3-90ac701486e8. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 961.792018] env[62600]: DEBUG oslo_concurrency.lockutils [req-1ce6075e-c335-4acf-892d-a83071c8c4de req-7b444df9-0bb1-43cc-b444-b4f7ab0d498f service nova] Acquiring lock "refresh_cache-4d12c133-84d7-4ce9-83db-e4b70b8947f3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.792018] env[62600]: DEBUG oslo_concurrency.lockutils [req-1ce6075e-c335-4acf-892d-a83071c8c4de req-7b444df9-0bb1-43cc-b444-b4f7ab0d498f service nova] Acquired lock "refresh_cache-4d12c133-84d7-4ce9-83db-e4b70b8947f3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.792018] env[62600]: DEBUG nova.network.neutron [req-1ce6075e-c335-4acf-892d-a83071c8c4de req-7b444df9-0bb1-43cc-b444-b4f7ab0d498f service nova] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Refreshing network info cache for port a3a87739-c27f-4a13-a4f3-90ac701486e8 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 961.796471] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222842, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.834506] env[62600]: DEBUG oslo_vmware.api [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Task: {'id': task-1222843, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.27705} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.834825] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 961.835035] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 961.835232] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 961.835412] env[62600]: INFO nova.compute.manager [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Took 1.12 seconds to destroy the instance on the hypervisor. [ 961.835689] env[62600]: DEBUG oslo.service.loopingcall [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.835898] env[62600]: DEBUG nova.compute.manager [-] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 961.835995] env[62600]: DEBUG nova.network.neutron [-] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 962.033262] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222844, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.035189] env[62600]: DEBUG oslo_concurrency.lockutils [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.995s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.038214] env[62600]: INFO nova.compute.manager [-] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Took 1.84 seconds to deallocate network for instance. [ 962.066864] env[62600]: INFO nova.scheduler.client.report [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleted allocations for instance f18df958-2885-4d8c-a9ad-140faaed1178 [ 962.068460] env[62600]: DEBUG nova.compute.manager [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 962.120351] env[62600]: DEBUG nova.virt.hardware [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 962.120605] env[62600]: DEBUG nova.virt.hardware [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 962.120769] env[62600]: DEBUG nova.virt.hardware [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 962.120953] env[62600]: DEBUG nova.virt.hardware [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 962.121158] env[62600]: DEBUG nova.virt.hardware [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 962.121335] env[62600]: DEBUG nova.virt.hardware [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 962.121570] env[62600]: DEBUG nova.virt.hardware [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 962.121760] env[62600]: DEBUG nova.virt.hardware [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 962.121978] env[62600]: DEBUG nova.virt.hardware [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 962.122176] env[62600]: DEBUG nova.virt.hardware [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 962.122355] env[62600]: DEBUG nova.virt.hardware [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 962.123346] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bd88d17-72c1-474e-8036-86411dcf6339 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.127099] env[62600]: DEBUG nova.network.neutron [-] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.134659] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01c688e-3e34-450c-915f-968cd1875cac {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.167056] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222845, 'name': CreateVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.176735] env[62600]: DEBUG nova.compute.manager [req-a37e12c0-d612-4976-9059-1eee33598557 req-4ae35cd4-9733-4ba4-9cd2-b5b4cc060127 service nova] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Received event network-vif-deleted-df39b522-a9a3-4d74-afe8-c68d5355ee10 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 962.178067] env[62600]: INFO nova.compute.manager [req-a37e12c0-d612-4976-9059-1eee33598557 req-4ae35cd4-9733-4ba4-9cd2-b5b4cc060127 service nova] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Neutron deleted interface df39b522-a9a3-4d74-afe8-c68d5355ee10; detaching it from the instance and deleting it from the info cache [ 962.178067] env[62600]: DEBUG nova.network.neutron [req-a37e12c0-d612-4976-9059-1eee33598557 req-4ae35cd4-9733-4ba4-9cd2-b5b4cc060127 service nova] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.295240] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222842, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.97604} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.295551] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 962.299691] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e817b880-5517-4b62-9de4-0ac44b815e37 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.323262] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 39deb498-6bf0-4f3b-932b-8068fc48271e/39deb498-6bf0-4f3b-932b-8068fc48271e.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 962.323262] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01a32c68-15b4-46ed-b338-04184bee039f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.342473] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 962.342473] env[62600]: value = "task-1222846" [ 962.342473] env[62600]: _type = "Task" [ 962.342473] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.352790] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222846, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.382139] env[62600]: DEBUG nova.network.neutron [-] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.534632] env[62600]: DEBUG oslo_vmware.api [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222844, 'name': PowerOnVM_Task, 'duration_secs': 0.787773} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.534987] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 962.535255] env[62600]: INFO nova.compute.manager [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Took 8.70 seconds to spawn the instance on the hypervisor. [ 962.535487] env[62600]: DEBUG nova.compute.manager [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 962.536386] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53203dd0-b111-4ed1-8e0f-5b91c23df776 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.544773] env[62600]: DEBUG oslo_concurrency.lockutils [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.545572] env[62600]: DEBUG oslo_concurrency.lockutils [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.545880] env[62600]: DEBUG nova.objects.instance [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lazy-loading 'resources' on Instance uuid 84dc284d-c7f4-4fc2-b539-8d4820429b02 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 962.583497] env[62600]: DEBUG oslo_concurrency.lockutils [None req-80aed8cd-d742-4786-a34c-d4d3be27f79b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "f18df958-2885-4d8c-a9ad-140faaed1178" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.192s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.629908] env[62600]: INFO nova.compute.manager [-] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Took 0.79 seconds to deallocate network for instance. [ 962.647249] env[62600]: DEBUG nova.network.neutron [req-1ce6075e-c335-4acf-892d-a83071c8c4de req-7b444df9-0bb1-43cc-b444-b4f7ab0d498f service nova] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Updated VIF entry in instance network info cache for port a3a87739-c27f-4a13-a4f3-90ac701486e8. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 962.648218] env[62600]: DEBUG nova.network.neutron [req-1ce6075e-c335-4acf-892d-a83071c8c4de req-7b444df9-0bb1-43cc-b444-b4f7ab0d498f service nova] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Updating instance_info_cache with network_info: [{"id": "a3a87739-c27f-4a13-a4f3-90ac701486e8", "address": "fa:16:3e:f8:77:26", "network": {"id": "1d2052cf-5be1-48e7-887a-18f367b7823f", "bridge": "br-int", "label": "tempest-ServersTestJSON-106099679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f91091f83ee4a2091507ca994e3d52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3a87739-c2", "ovs_interfaceid": "a3a87739-c27f-4a13-a4f3-90ac701486e8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.669443] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222845, 'name': CreateVM_Task, 'duration_secs': 0.59971} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.669731] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 962.670395] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.670634] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.671112] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 962.671753] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a9b72d1-9bed-446e-aa25-512bbb41b938 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.677499] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 962.677499] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521c02f0-297c-0dc8-5c9d-3a27a7546a32" [ 962.677499] env[62600]: _type = "Task" [ 962.677499] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.681070] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-00c83659-5696-46b0-8b89-c65babf3b250 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.688326] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521c02f0-297c-0dc8-5c9d-3a27a7546a32, 'name': SearchDatastore_Task, 'duration_secs': 0.008991} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.689337] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.689575] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 962.689805] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.689959] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.690181] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 962.690683] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bbdc9786-2682-47d7-afa3-19896b4ea065 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.694854] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2114630-89db-4ff9-9d40-e1de06721b02 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.712792] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 962.712987] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 962.713734] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0d413b5-bf77-4f4b-8ac1-d49238c0d0b0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.726206] env[62600]: DEBUG nova.compute.manager [req-a37e12c0-d612-4976-9059-1eee33598557 req-4ae35cd4-9733-4ba4-9cd2-b5b4cc060127 service nova] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Detach interface failed, port_id=df39b522-a9a3-4d74-afe8-c68d5355ee10, reason: Instance 81333628-2616-4557-88d1-0e7164e1b16d could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 962.729683] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 962.729683] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52dc5150-4b49-3f67-2d34-fb78caad5435" [ 962.729683] env[62600]: _type = "Task" [ 962.729683] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.738314] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52dc5150-4b49-3f67-2d34-fb78caad5435, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.739463] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Volume attach. Driver type: vmdk {{(pid=62600) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 962.739717] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264349', 'volume_id': 'e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf', 'name': 'volume-e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '248a5371-6ff7-4da7-ae97-d638bea0123a', 'attached_at': '', 'detached_at': '', 'volume_id': 'e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf', 'serial': 'e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf'} {{(pid=62600) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 962.741042] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd33954e-0e50-4bae-9768-a3551bea638a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.759495] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e249b2-6d8d-4b0c-a514-36f5406970b9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.784415] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] volume-e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf/volume-e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 962.784415] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1fec63ad-bc02-4499-98fc-54854b5763cb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.797106] env[62600]: DEBUG nova.network.neutron [-] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.803969] env[62600]: DEBUG oslo_vmware.api [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 962.803969] env[62600]: value = "task-1222847" [ 962.803969] env[62600]: _type = "Task" [ 962.803969] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.814266] env[62600]: DEBUG oslo_vmware.api [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222847, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.852432] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222846, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.887167] env[62600]: INFO nova.compute.manager [-] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Took 1.30 seconds to deallocate network for instance. [ 963.056441] env[62600]: INFO nova.compute.manager [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Took 24.86 seconds to build instance. [ 963.138152] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.151064] env[62600]: DEBUG oslo_concurrency.lockutils [req-1ce6075e-c335-4acf-892d-a83071c8c4de req-7b444df9-0bb1-43cc-b444-b4f7ab0d498f service nova] Releasing lock "refresh_cache-4d12c133-84d7-4ce9-83db-e4b70b8947f3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.151363] env[62600]: DEBUG nova.compute.manager [req-1ce6075e-c335-4acf-892d-a83071c8c4de req-7b444df9-0bb1-43cc-b444-b4f7ab0d498f service nova] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Received event network-vif-deleted-64a8ba35-fe05-4c77-ac73-f420bd31a46d {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.245215] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52dc5150-4b49-3f67-2d34-fb78caad5435, 'name': SearchDatastore_Task, 'duration_secs': 0.008695} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.246609] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d12298f-15e3-4f33-b179-00a671beb856 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.253679] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f5ca212-5811-40fa-8a37-f6f990a04a89 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.258928] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 963.258928] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5212e76b-a6f8-2693-9d4f-5c9962d8b764" [ 963.258928] env[62600]: _type = "Task" [ 963.258928] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.266342] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6f3b346-2db6-4349-a41a-9e965fc47930 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.277932] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5212e76b-a6f8-2693-9d4f-5c9962d8b764, 'name': SearchDatastore_Task, 'duration_secs': 0.012753} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.278577] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.278848] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 4d12c133-84d7-4ce9-83db-e4b70b8947f3/4d12c133-84d7-4ce9-83db-e4b70b8947f3.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 963.279129] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2e724299-48f2-45f8-9953-875702d4be13 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.309130] env[62600]: INFO nova.compute.manager [-] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Took 1.60 seconds to deallocate network for instance. [ 963.315390] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95fcc23e-f4bb-4629-89ef-e6d0c89deb9f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.323124] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 963.323124] env[62600]: value = "task-1222848" [ 963.323124] env[62600]: _type = "Task" [ 963.323124] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.331337] env[62600]: DEBUG oslo_vmware.api [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.332874] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb75145f-5f0a-4dae-bff9-1b192c23c84a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.340161] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222848, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.349582] env[62600]: DEBUG nova.compute.provider_tree [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.362339] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222846, 'name': ReconfigVM_Task, 'duration_secs': 0.579718} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.362708] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 39deb498-6bf0-4f3b-932b-8068fc48271e/39deb498-6bf0-4f3b-932b-8068fc48271e.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 963.363397] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-747b6df6-5096-4d0b-bc5f-3af84fbd300a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.370867] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 963.370867] env[62600]: value = "task-1222849" [ 963.370867] env[62600]: _type = "Task" [ 963.370867] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.382805] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222849, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.395348] env[62600]: DEBUG oslo_concurrency.lockutils [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.436239] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8acfcf0-99e8-488d-b68b-2f947d4fd449 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.444644] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a81487a1-b1f6-481e-bb85-a0044b69442b tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Suspending the VM {{(pid=62600) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 963.444945] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-5865f0b5-6d77-40a6-bf2c-70a6ea2eac76 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.451399] env[62600]: DEBUG oslo_vmware.api [None req-a81487a1-b1f6-481e-bb85-a0044b69442b tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 963.451399] env[62600]: value = "task-1222850" [ 963.451399] env[62600]: _type = "Task" [ 963.451399] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.459741] env[62600]: DEBUG oslo_vmware.api [None req-a81487a1-b1f6-481e-bb85-a0044b69442b tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222850, 'name': SuspendVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.558861] env[62600]: DEBUG oslo_concurrency.lockutils [None req-6f23c585-f2ae-43f3-becb-97e885484a98 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "234dc7f8-e2ef-4f20-b25b-dcd61ce20c57" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.376s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.761893] env[62600]: DEBUG nova.network.neutron [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Successfully updated port: 3753b6a6-f27d-4e1d-b801-91324a7e4e0e {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 963.825589] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.826786] env[62600]: DEBUG oslo_vmware.api [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222847, 'name': ReconfigVM_Task, 'duration_secs': 0.664147} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.832902] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Reconfigured VM instance instance-00000050 to attach disk [datastore1] volume-e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf/volume-e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 963.839864] env[62600]: DEBUG nova.compute.manager [req-95cdee7d-4d0f-4e25-a02a-17482661fb3f req-aeb5f5fe-5880-47fa-989b-cdad5aa576b6 service nova] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Received event network-vif-deleted-2bf1b736-051a-4ca6-8800-a34547f63394 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.840122] env[62600]: DEBUG nova.compute.manager [req-95cdee7d-4d0f-4e25-a02a-17482661fb3f req-aeb5f5fe-5880-47fa-989b-cdad5aa576b6 service nova] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Received event network-vif-plugged-3753b6a6-f27d-4e1d-b801-91324a7e4e0e {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.840389] env[62600]: DEBUG oslo_concurrency.lockutils [req-95cdee7d-4d0f-4e25-a02a-17482661fb3f req-aeb5f5fe-5880-47fa-989b-cdad5aa576b6 service nova] Acquiring lock "02c6dac9-535b-436b-bbee-c8031cb5d7f5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.840629] env[62600]: DEBUG oslo_concurrency.lockutils [req-95cdee7d-4d0f-4e25-a02a-17482661fb3f req-aeb5f5fe-5880-47fa-989b-cdad5aa576b6 service nova] Lock "02c6dac9-535b-436b-bbee-c8031cb5d7f5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.840819] env[62600]: DEBUG oslo_concurrency.lockutils [req-95cdee7d-4d0f-4e25-a02a-17482661fb3f req-aeb5f5fe-5880-47fa-989b-cdad5aa576b6 service nova] Lock "02c6dac9-535b-436b-bbee-c8031cb5d7f5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.841014] env[62600]: DEBUG nova.compute.manager [req-95cdee7d-4d0f-4e25-a02a-17482661fb3f req-aeb5f5fe-5880-47fa-989b-cdad5aa576b6 service nova] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] No waiting events found dispatching network-vif-plugged-3753b6a6-f27d-4e1d-b801-91324a7e4e0e {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 963.841202] env[62600]: WARNING nova.compute.manager [req-95cdee7d-4d0f-4e25-a02a-17482661fb3f req-aeb5f5fe-5880-47fa-989b-cdad5aa576b6 service nova] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Received unexpected event network-vif-plugged-3753b6a6-f27d-4e1d-b801-91324a7e4e0e for instance with vm_state building and task_state spawning. [ 963.841453] env[62600]: DEBUG nova.compute.manager [req-95cdee7d-4d0f-4e25-a02a-17482661fb3f req-aeb5f5fe-5880-47fa-989b-cdad5aa576b6 service nova] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Received event network-changed-3753b6a6-f27d-4e1d-b801-91324a7e4e0e {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.841715] env[62600]: DEBUG nova.compute.manager [req-95cdee7d-4d0f-4e25-a02a-17482661fb3f req-aeb5f5fe-5880-47fa-989b-cdad5aa576b6 service nova] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Refreshing instance network info cache due to event network-changed-3753b6a6-f27d-4e1d-b801-91324a7e4e0e. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 963.842067] env[62600]: DEBUG oslo_concurrency.lockutils [req-95cdee7d-4d0f-4e25-a02a-17482661fb3f req-aeb5f5fe-5880-47fa-989b-cdad5aa576b6 service nova] Acquiring lock "refresh_cache-02c6dac9-535b-436b-bbee-c8031cb5d7f5" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.842277] env[62600]: DEBUG oslo_concurrency.lockutils [req-95cdee7d-4d0f-4e25-a02a-17482661fb3f req-aeb5f5fe-5880-47fa-989b-cdad5aa576b6 service nova] Acquired lock "refresh_cache-02c6dac9-535b-436b-bbee-c8031cb5d7f5" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.842541] env[62600]: DEBUG nova.network.neutron [req-95cdee7d-4d0f-4e25-a02a-17482661fb3f req-aeb5f5fe-5880-47fa-989b-cdad5aa576b6 service nova] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Refreshing network info cache for port 3753b6a6-f27d-4e1d-b801-91324a7e4e0e {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 963.843913] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-966372f9-9d6d-4036-b5da-f5554bdf3b9d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.859457] env[62600]: DEBUG nova.scheduler.client.report [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 963.867037] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222848, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.535986} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.867037] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 4d12c133-84d7-4ce9-83db-e4b70b8947f3/4d12c133-84d7-4ce9-83db-e4b70b8947f3.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 963.867037] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 963.867370] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f1fc0d03-b9b7-441b-8940-2c46789f1dea {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.870518] env[62600]: DEBUG oslo_vmware.api [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 963.870518] env[62600]: value = "task-1222851" [ 963.870518] env[62600]: _type = "Task" [ 963.870518] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.879563] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 963.879563] env[62600]: value = "task-1222852" [ 963.879563] env[62600]: _type = "Task" [ 963.879563] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.883657] env[62600]: DEBUG oslo_vmware.api [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222851, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.893042] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222849, 'name': Rename_Task, 'duration_secs': 0.148517} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.893772] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 963.894190] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-298c3039-564b-4207-8697-1976dc23a8d2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.899298] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222852, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.904293] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 963.904293] env[62600]: value = "task-1222853" [ 963.904293] env[62600]: _type = "Task" [ 963.904293] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.908342] env[62600]: DEBUG nova.network.neutron [req-95cdee7d-4d0f-4e25-a02a-17482661fb3f req-aeb5f5fe-5880-47fa-989b-cdad5aa576b6 service nova] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 963.913542] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222853, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.962216] env[62600]: DEBUG oslo_vmware.api [None req-a81487a1-b1f6-481e-bb85-a0044b69442b tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222850, 'name': SuspendVM_Task} progress is 50%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.029718] env[62600]: DEBUG nova.network.neutron [req-95cdee7d-4d0f-4e25-a02a-17482661fb3f req-aeb5f5fe-5880-47fa-989b-cdad5aa576b6 service nova] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.233010] env[62600]: DEBUG nova.compute.manager [req-11a31a9d-208a-4d2e-94f4-feda3585a630 req-db554bf3-9bd7-44eb-aced-abb1df93f86c service nova] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Received event network-vif-deleted-07d47e0c-0331-4404-9aa9-3568372c18ff {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.265738] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "refresh_cache-02c6dac9-535b-436b-bbee-c8031cb5d7f5" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.369288] env[62600]: DEBUG oslo_concurrency.lockutils [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.822s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.370077] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.232s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.370490] env[62600]: DEBUG nova.objects.instance [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Lazy-loading 'resources' on Instance uuid 9ed08410-a4cb-43c4-9271-2c04d9c87eac {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.383252] env[62600]: DEBUG oslo_vmware.api [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222851, 'name': ReconfigVM_Task, 'duration_secs': 0.157749} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.384951] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264349', 'volume_id': 'e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf', 'name': 'volume-e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '248a5371-6ff7-4da7-ae97-d638bea0123a', 'attached_at': '', 'detached_at': '', 'volume_id': 'e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf', 'serial': 'e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf'} {{(pid=62600) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 964.396177] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222852, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068088} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.396846] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 964.397803] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e59c11a-592c-47b3-87da-90807c444046 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.403200] env[62600]: INFO nova.scheduler.client.report [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Deleted allocations for instance 84dc284d-c7f4-4fc2-b539-8d4820429b02 [ 964.434207] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] 4d12c133-84d7-4ce9-83db-e4b70b8947f3/4d12c133-84d7-4ce9-83db-e4b70b8947f3.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 964.434384] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-945d39ea-880e-413d-b0bc-077911adf62b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.454824] env[62600]: DEBUG oslo_vmware.api [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222853, 'name': PowerOnVM_Task, 'duration_secs': 0.51086} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.458310] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 964.458631] env[62600]: INFO nova.compute.manager [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Took 8.25 seconds to spawn the instance on the hypervisor. [ 964.458994] env[62600]: DEBUG nova.compute.manager [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.461029] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db72f5a-2821-4bc7-afc0-af18a2156584 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.463545] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 964.463545] env[62600]: value = "task-1222854" [ 964.463545] env[62600]: _type = "Task" [ 964.463545] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.472513] env[62600]: DEBUG oslo_vmware.api [None req-a81487a1-b1f6-481e-bb85-a0044b69442b tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222850, 'name': SuspendVM_Task, 'duration_secs': 0.724552} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.477152] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a81487a1-b1f6-481e-bb85-a0044b69442b tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Suspended the VM {{(pid=62600) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 964.477498] env[62600]: DEBUG nova.compute.manager [None req-a81487a1-b1f6-481e-bb85-a0044b69442b tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.478946] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-539d0d52-069f-422e-9152-7eeefe8f279f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.486351] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222854, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.532961] env[62600]: DEBUG oslo_concurrency.lockutils [req-95cdee7d-4d0f-4e25-a02a-17482661fb3f req-aeb5f5fe-5880-47fa-989b-cdad5aa576b6 service nova] Releasing lock "refresh_cache-02c6dac9-535b-436b-bbee-c8031cb5d7f5" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.532961] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "refresh_cache-02c6dac9-535b-436b-bbee-c8031cb5d7f5" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.533081] env[62600]: DEBUG nova.network.neutron [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 964.930119] env[62600]: DEBUG oslo_concurrency.lockutils [None req-113e6357-07d9-487a-b24d-4c1bc32e1948 tempest-ImagesTestJSON-1540326467 tempest-ImagesTestJSON-1540326467-project-member] Lock "84dc284d-c7f4-4fc2-b539-8d4820429b02" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.867s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.987766] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222854, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.996391] env[62600]: INFO nova.compute.manager [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Took 25.00 seconds to build instance. [ 965.083208] env[62600]: DEBUG nova.network.neutron [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 965.134530] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfe9c8b2-82a8-4e1a-b0b4-75c113958f80 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.145524] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dad2ecca-2345-4847-9ba8-823dfcc083c7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.192363] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4f2667-f60b-4720-a3ca-19d6d5fdd880 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.200487] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dccebac-c452-4ac7-8ed1-327d1b0072fb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.221012] env[62600]: DEBUG nova.compute.provider_tree [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.400796] env[62600]: DEBUG nova.network.neutron [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Updating instance_info_cache with network_info: [{"id": "3753b6a6-f27d-4e1d-b801-91324a7e4e0e", "address": "fa:16:3e:fd:6e:a3", "network": {"id": "95818650-8227-4d11-9d14-007f06c07790", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1738152115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73e664d178f7484a9f4741b4d9450e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3753b6a6-f2", "ovs_interfaceid": "3753b6a6-f27d-4e1d-b801-91324a7e4e0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.436799] env[62600]: DEBUG nova.objects.instance [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lazy-loading 'flavor' on Instance uuid 248a5371-6ff7-4da7-ae97-d638bea0123a {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.474907] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222854, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.501419] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2790ece9-ce80-48cd-a508-98d0a34aecf2 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "39deb498-6bf0-4f3b-932b-8068fc48271e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.515s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.709398] env[62600]: INFO nova.compute.manager [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Rescuing [ 965.709966] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.710377] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquired lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.710377] env[62600]: DEBUG nova.network.neutron [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 965.722062] env[62600]: DEBUG nova.scheduler.client.report [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 965.904045] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "refresh_cache-02c6dac9-535b-436b-bbee-c8031cb5d7f5" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.904537] env[62600]: DEBUG nova.compute.manager [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Instance network_info: |[{"id": "3753b6a6-f27d-4e1d-b801-91324a7e4e0e", "address": "fa:16:3e:fd:6e:a3", "network": {"id": "95818650-8227-4d11-9d14-007f06c07790", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1738152115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73e664d178f7484a9f4741b4d9450e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3753b6a6-f2", "ovs_interfaceid": "3753b6a6-f27d-4e1d-b801-91324a7e4e0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 965.904934] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fd:6e:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43798f54-0c07-4417-a23f-58bb6b7e204b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3753b6a6-f27d-4e1d-b801-91324a7e4e0e', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 965.914917] env[62600]: DEBUG oslo.service.loopingcall [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.915334] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 965.915636] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bde49754-d011-4ca5-9712-88bb20cb63d1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.943395] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 965.943395] env[62600]: value = "task-1222855" [ 965.943395] env[62600]: _type = "Task" [ 965.943395] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.947214] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1ea23abc-8ca5-418a-a071-c4263b20c214 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "248a5371-6ff7-4da7-ae97-d638bea0123a" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.852s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.953986] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222855, 'name': CreateVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.977377] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222854, 'name': ReconfigVM_Task, 'duration_secs': 1.304176} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.978297] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Reconfigured VM instance instance-00000059 to attach disk [datastore1] 4d12c133-84d7-4ce9-83db-e4b70b8947f3/4d12c133-84d7-4ce9-83db-e4b70b8947f3.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 965.978949] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b8abe044-2cfa-497d-94b6-4d978216a33e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.985949] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 965.985949] env[62600]: value = "task-1222856" [ 965.985949] env[62600]: _type = "Task" [ 965.985949] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.995397] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222856, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.084214] env[62600]: DEBUG oslo_concurrency.lockutils [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "234dc7f8-e2ef-4f20-b25b-dcd61ce20c57" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.084214] env[62600]: DEBUG oslo_concurrency.lockutils [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "234dc7f8-e2ef-4f20-b25b-dcd61ce20c57" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.084214] env[62600]: DEBUG oslo_concurrency.lockutils [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "234dc7f8-e2ef-4f20-b25b-dcd61ce20c57-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.084214] env[62600]: DEBUG oslo_concurrency.lockutils [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "234dc7f8-e2ef-4f20-b25b-dcd61ce20c57-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.084214] env[62600]: DEBUG oslo_concurrency.lockutils [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "234dc7f8-e2ef-4f20-b25b-dcd61ce20c57-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.085996] env[62600]: INFO nova.compute.manager [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Terminating instance [ 966.087945] env[62600]: DEBUG nova.compute.manager [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 966.088274] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 966.089244] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a638f3b8-33af-4204-9168-3fb7615231fb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.100240] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 966.100240] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2357bc1c-e98d-4b03-8eb8-1d6b01e10fa6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.195345] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 966.195733] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 966.196156] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Deleting the datastore file [datastore1] 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 966.196415] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2db0de32-93c3-4386-b86f-e9190a5b7ac4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.203281] env[62600]: DEBUG oslo_vmware.api [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 966.203281] env[62600]: value = "task-1222858" [ 966.203281] env[62600]: _type = "Task" [ 966.203281] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.217335] env[62600]: DEBUG oslo_vmware.api [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222858, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.227868] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.858s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.230492] env[62600]: DEBUG oslo_concurrency.lockutils [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.835s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.230664] env[62600]: DEBUG nova.objects.instance [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lazy-loading 'resources' on Instance uuid 81333628-2616-4557-88d1-0e7164e1b16d {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.276598] env[62600]: DEBUG nova.compute.manager [req-9ca75d41-026f-43b8-b622-de77e32ef817 req-d08de249-47ce-48ee-a264-0d03db9498f8 service nova] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Received event network-changed-ead68533-9f71-4ab4-a47d-af3e0c397bd4 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.276798] env[62600]: DEBUG nova.compute.manager [req-9ca75d41-026f-43b8-b622-de77e32ef817 req-d08de249-47ce-48ee-a264-0d03db9498f8 service nova] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Refreshing instance network info cache due to event network-changed-ead68533-9f71-4ab4-a47d-af3e0c397bd4. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 966.277022] env[62600]: DEBUG oslo_concurrency.lockutils [req-9ca75d41-026f-43b8-b622-de77e32ef817 req-d08de249-47ce-48ee-a264-0d03db9498f8 service nova] Acquiring lock "refresh_cache-39deb498-6bf0-4f3b-932b-8068fc48271e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.277713] env[62600]: DEBUG oslo_concurrency.lockutils [req-9ca75d41-026f-43b8-b622-de77e32ef817 req-d08de249-47ce-48ee-a264-0d03db9498f8 service nova] Acquired lock "refresh_cache-39deb498-6bf0-4f3b-932b-8068fc48271e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.277713] env[62600]: DEBUG nova.network.neutron [req-9ca75d41-026f-43b8-b622-de77e32ef817 req-d08de249-47ce-48ee-a264-0d03db9498f8 service nova] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Refreshing network info cache for port ead68533-9f71-4ab4-a47d-af3e0c397bd4 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 966.279696] env[62600]: INFO nova.scheduler.client.report [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Deleted allocations for instance 9ed08410-a4cb-43c4-9271-2c04d9c87eac [ 966.453715] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222855, 'name': CreateVM_Task, 'duration_secs': 0.343013} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.453940] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 966.454914] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.455140] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.455672] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 966.455970] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f09354b7-f372-42cc-9e17-c09a065680ad {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.463459] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 966.463459] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5285ced4-393b-5709-7a84-92f0b408c92b" [ 966.463459] env[62600]: _type = "Task" [ 966.463459] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.470922] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5285ced4-393b-5709-7a84-92f0b408c92b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.494755] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222856, 'name': Rename_Task, 'duration_secs': 0.141422} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.495044] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 966.495298] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-92b37d7b-6360-46db-a70b-8a7201acb669 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.501956] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 966.501956] env[62600]: value = "task-1222859" [ 966.501956] env[62600]: _type = "Task" [ 966.501956] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.509690] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222859, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.544158] env[62600]: DEBUG nova.network.neutron [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Updating instance_info_cache with network_info: [{"id": "32f5d56d-a029-4469-ad3c-77a4ca51bc02", "address": "fa:16:3e:da:8c:98", "network": {"id": "036dc114-6ea0-46cb-a958-4a6137e7e85f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-728627542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e1772c9a6c4671ab89df2e8b40cae1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32f5d56d-a0", "ovs_interfaceid": "32f5d56d-a029-4469-ad3c-77a4ca51bc02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.714007] env[62600]: DEBUG oslo_vmware.api [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222858, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160826} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.714290] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 966.714497] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 966.714703] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 966.714905] env[62600]: INFO nova.compute.manager [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Took 0.63 seconds to destroy the instance on the hypervisor. [ 966.715172] env[62600]: DEBUG oslo.service.loopingcall [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 966.715374] env[62600]: DEBUG nova.compute.manager [-] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 966.715472] env[62600]: DEBUG nova.network.neutron [-] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 966.796882] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7306a450-2eb0-4798-b63c-e1d63d575153 tempest-ImagesOneServerTestJSON-1257150813 tempest-ImagesOneServerTestJSON-1257150813-project-member] Lock "9ed08410-a4cb-43c4-9271-2c04d9c87eac" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.092s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.980631] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5285ced4-393b-5709-7a84-92f0b408c92b, 'name': SearchDatastore_Task, 'duration_secs': 0.009292} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.982135] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.982386] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 966.982624] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.982769] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.983086] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 966.983780] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f5666f-b7c5-4680-b931-c18dc415dc27 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.990019] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0bb90b13-4c28-489d-a082-b645f6e29a67 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.995289] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02e8d200-8a5b-4b2d-bc33-5141c86c7131 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.000586] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 967.003627] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 967.003627] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68e88f0d-ce3e-4231-9e68-fc5944c9bd79 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.037863] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9953ada-7079-43ca-86e0-5037e77ccada {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.043141] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 967.043141] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5282a83c-65a3-90c9-7d16-93ac1bc59c1e" [ 967.043141] env[62600]: _type = "Task" [ 967.043141] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.052441] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Releasing lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.054962] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222859, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.057236] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8afe99-5e5b-4422-bcfb-c97ba1744e9d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.066132] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5282a83c-65a3-90c9-7d16-93ac1bc59c1e, 'name': SearchDatastore_Task, 'duration_secs': 0.011241} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.069472] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e4d44e4-6aa5-4ae9-ac16-ff784ac46260 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.079905] env[62600]: DEBUG nova.compute.provider_tree [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.086442] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 967.086442] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52686822-c3d2-1729-832d-b2ce29e7e8a0" [ 967.086442] env[62600]: _type = "Task" [ 967.086442] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.103982] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52686822-c3d2-1729-832d-b2ce29e7e8a0, 'name': SearchDatastore_Task, 'duration_secs': 0.010879} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.104136] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.104392] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 02c6dac9-535b-436b-bbee-c8031cb5d7f5/02c6dac9-535b-436b-bbee-c8031cb5d7f5.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 967.104729] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aada6c7d-b33d-4da8-a14d-ffa901c89b83 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.114497] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 967.114497] env[62600]: value = "task-1222860" [ 967.114497] env[62600]: _type = "Task" [ 967.114497] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.121149] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222860, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.277008] env[62600]: DEBUG nova.compute.manager [req-fc48a7b6-56ba-4980-9258-6e2f01bfa046 req-7a4c060c-3883-4d85-b331-3323c0a34d29 service nova] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Received event network-vif-deleted-80551113-7315-4224-ac50-2f6b14d0a6b0 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.277234] env[62600]: INFO nova.compute.manager [req-fc48a7b6-56ba-4980-9258-6e2f01bfa046 req-7a4c060c-3883-4d85-b331-3323c0a34d29 service nova] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Neutron deleted interface 80551113-7315-4224-ac50-2f6b14d0a6b0; detaching it from the instance and deleting it from the info cache [ 967.277420] env[62600]: DEBUG nova.network.neutron [req-fc48a7b6-56ba-4980-9258-6e2f01bfa046 req-7a4c060c-3883-4d85-b331-3323c0a34d29 service nova] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.347152] env[62600]: DEBUG nova.network.neutron [req-9ca75d41-026f-43b8-b622-de77e32ef817 req-d08de249-47ce-48ee-a264-0d03db9498f8 service nova] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Updated VIF entry in instance network info cache for port ead68533-9f71-4ab4-a47d-af3e0c397bd4. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 967.347152] env[62600]: DEBUG nova.network.neutron [req-9ca75d41-026f-43b8-b622-de77e32ef817 req-d08de249-47ce-48ee-a264-0d03db9498f8 service nova] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Updating instance_info_cache with network_info: [{"id": "ead68533-9f71-4ab4-a47d-af3e0c397bd4", "address": "fa:16:3e:f6:16:4c", "network": {"id": "e7547ac4-24a6-4778-a17a-40e79e7a0d0c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-320671681-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.133", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3bb131f3994f7a8673dafff6d353c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapead68533-9f", "ovs_interfaceid": "ead68533-9f71-4ab4-a47d-af3e0c397bd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.542820] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222859, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.587060] env[62600]: DEBUG nova.scheduler.client.report [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 967.603023] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 967.603023] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e6aefc4-9fbb-4288-a7fe-1d2f383dcef9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.614027] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 967.614027] env[62600]: value = "task-1222861" [ 967.614027] env[62600]: _type = "Task" [ 967.614027] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.623884] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222860, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.629209] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222861, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.728777] env[62600]: DEBUG nova.network.neutron [-] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.779780] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bfcc1e9c-92ba-4a20-b851-77c87228095f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.790043] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd8ac8d-e5f2-4791-83e6-b5e7843cdbd5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.821186] env[62600]: DEBUG nova.compute.manager [req-fc48a7b6-56ba-4980-9258-6e2f01bfa046 req-7a4c060c-3883-4d85-b331-3323c0a34d29 service nova] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Detach interface failed, port_id=80551113-7315-4224-ac50-2f6b14d0a6b0, reason: Instance 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 967.848720] env[62600]: DEBUG oslo_concurrency.lockutils [req-9ca75d41-026f-43b8-b622-de77e32ef817 req-d08de249-47ce-48ee-a264-0d03db9498f8 service nova] Releasing lock "refresh_cache-39deb498-6bf0-4f3b-932b-8068fc48271e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.044165] env[62600]: DEBUG oslo_vmware.api [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222859, 'name': PowerOnVM_Task, 'duration_secs': 1.129491} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.044521] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 968.044706] env[62600]: INFO nova.compute.manager [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Took 9.41 seconds to spawn the instance on the hypervisor. [ 968.044907] env[62600]: DEBUG nova.compute.manager [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 968.045713] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf3a1f10-b3ab-47a7-9fbf-8864cd554be9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.090673] env[62600]: DEBUG oslo_concurrency.lockutils [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.860s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.093264] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.268s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.093414] env[62600]: DEBUG nova.objects.instance [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lazy-loading 'resources' on Instance uuid 784e6d19-36ef-4c01-9e9f-5a083ed08608 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.121317] env[62600]: INFO nova.scheduler.client.report [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Deleted allocations for instance 81333628-2616-4557-88d1-0e7164e1b16d [ 968.126600] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222861, 'name': PowerOffVM_Task, 'duration_secs': 0.333678} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.130406] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 968.138242] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9df07a4a-95a4-451c-83ba-2067271e4226 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.138242] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222860, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.530988} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.138242] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 02c6dac9-535b-436b-bbee-c8031cb5d7f5/02c6dac9-535b-436b-bbee-c8031cb5d7f5.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 968.138242] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 968.139081] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9f45a7da-73b1-441b-8610-e1bc0592ee74 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.168246] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b33452-c054-44ac-8902-e2ec92e29a5f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.171720] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 968.171720] env[62600]: value = "task-1222862" [ 968.171720] env[62600]: _type = "Task" [ 968.171720] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.187101] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222862, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.217033] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 968.217033] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1ad8107d-883f-4fde-bbfa-f49e0f8ebac7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.228018] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 968.228018] env[62600]: value = "task-1222863" [ 968.228018] env[62600]: _type = "Task" [ 968.228018] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.230909] env[62600]: INFO nova.compute.manager [-] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Took 1.52 seconds to deallocate network for instance. [ 968.240602] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] VM already powered off {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 968.240973] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 968.241364] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.241760] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.242078] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 968.244830] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-64b0a6d8-8e04-4381-abf9-020adaa7f8c4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.253019] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 968.253019] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 968.253019] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a2dd264-8ca7-4c9a-bd55-e20e5bc3cff4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.257453] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 968.257453] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]523c634f-47a1-23a8-88c6-48ca1325bbad" [ 968.257453] env[62600]: _type = "Task" [ 968.257453] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.265581] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]523c634f-47a1-23a8-88c6-48ca1325bbad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.571135] env[62600]: INFO nova.compute.manager [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Took 22.79 seconds to build instance. [ 968.637110] env[62600]: DEBUG oslo_concurrency.lockutils [None req-bcf6078a-af46-45bd-bac3-3c07596578cb tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "81333628-2616-4557-88d1-0e7164e1b16d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.259s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.685284] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222862, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.211753} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.687799] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 968.689347] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-476b059d-72cb-405b-8009-9754b29096f7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.722606] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 02c6dac9-535b-436b-bbee-c8031cb5d7f5/02c6dac9-535b-436b-bbee-c8031cb5d7f5.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 968.723519] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76b11260-2633-4822-8bdf-0dfba67f9b46 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.743452] env[62600]: DEBUG oslo_concurrency.lockutils [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.749564] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 968.749564] env[62600]: value = "task-1222864" [ 968.749564] env[62600]: _type = "Task" [ 968.749564] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.759426] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222864, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.769576] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]523c634f-47a1-23a8-88c6-48ca1325bbad, 'name': SearchDatastore_Task, 'duration_secs': 0.009879} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.771313] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59455c12-4ac8-4907-a6f8-88177fb64913 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.776599] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 968.776599] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52bcd8e5-bf32-ee4b-dffc-99b13354cc99" [ 968.776599] env[62600]: _type = "Task" [ 968.776599] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.785382] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52bcd8e5-bf32-ee4b-dffc-99b13354cc99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.903980] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6cda322-f88a-4a82-94e6-43f9c43c799c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.912379] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c996f4-09e4-4e65-91b2-5d59f81d1a91 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.947466] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-205359eb-277c-401c-b2dc-1be3ff526cdd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.955127] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d654db8-f4a1-4339-b990-fc8b3db2cf80 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.969458] env[62600]: DEBUG nova.compute.provider_tree [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.073207] env[62600]: DEBUG oslo_concurrency.lockutils [None req-df62515e-d9e4-46d5-a651-3790079439d2 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "4d12c133-84d7-4ce9-83db-e4b70b8947f3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.306s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.266147] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222864, 'name': ReconfigVM_Task, 'duration_secs': 0.337808} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.266882] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 02c6dac9-535b-436b-bbee-c8031cb5d7f5/02c6dac9-535b-436b-bbee-c8031cb5d7f5.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 969.268145] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cdedf8cd-df95-4a5e-9573-afeaaf80ac0b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.275121] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 969.275121] env[62600]: value = "task-1222865" [ 969.275121] env[62600]: _type = "Task" [ 969.275121] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.291044] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52bcd8e5-bf32-ee4b-dffc-99b13354cc99, 'name': SearchDatastore_Task, 'duration_secs': 0.009876} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.294765] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.295061] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 248a5371-6ff7-4da7-ae97-d638bea0123a/e7bd1d77-bfff-4684-9545-f6a1c69efe58-rescue.vmdk. {{(pid=62600) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 969.298557] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222865, 'name': Rename_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.298557] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0c63bf8f-70c9-4cf6-a94f-0c9a896f9b5e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.302395] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 969.302395] env[62600]: value = "task-1222866" [ 969.302395] env[62600]: _type = "Task" [ 969.302395] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.311667] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222866, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.474130] env[62600]: DEBUG nova.scheduler.client.report [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 969.785510] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222865, 'name': Rename_Task, 'duration_secs': 0.145797} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.785793] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 969.786716] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-936655f5-8b4c-4e3d-a93d-ae0d4273604f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.793344] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 969.793344] env[62600]: value = "task-1222867" [ 969.793344] env[62600]: _type = "Task" [ 969.793344] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.801034] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222867, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.814167] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222866, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.47208} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.814752] env[62600]: INFO nova.virt.vmwareapi.ds_util [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 248a5371-6ff7-4da7-ae97-d638bea0123a/e7bd1d77-bfff-4684-9545-f6a1c69efe58-rescue.vmdk. [ 969.815813] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-facf6b96-5fe9-468e-ba8f-d84fe6c32c94 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.846612] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 248a5371-6ff7-4da7-ae97-d638bea0123a/e7bd1d77-bfff-4684-9545-f6a1c69efe58-rescue.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 969.846979] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a3a52207-0b55-4c57-a83d-e688d317f704 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.865868] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 969.865868] env[62600]: value = "task-1222868" [ 969.865868] env[62600]: _type = "Task" [ 969.865868] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.874156] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222868, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.953993] env[62600]: DEBUG oslo_concurrency.lockutils [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "4d12c133-84d7-4ce9-83db-e4b70b8947f3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.954306] env[62600]: DEBUG oslo_concurrency.lockutils [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "4d12c133-84d7-4ce9-83db-e4b70b8947f3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.954652] env[62600]: DEBUG oslo_concurrency.lockutils [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "4d12c133-84d7-4ce9-83db-e4b70b8947f3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.954768] env[62600]: DEBUG oslo_concurrency.lockutils [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "4d12c133-84d7-4ce9-83db-e4b70b8947f3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.955189] env[62600]: DEBUG oslo_concurrency.lockutils [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "4d12c133-84d7-4ce9-83db-e4b70b8947f3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.959369] env[62600]: INFO nova.compute.manager [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Terminating instance [ 969.961195] env[62600]: DEBUG nova.compute.manager [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 969.961301] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 969.962542] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a95d2f-a3ea-4c19-9185-09d1fd6b1039 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.970778] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 969.971072] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ea4140b-75fb-4f67-bb9f-fbd7e4c78591 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.977403] env[62600]: DEBUG oslo_vmware.api [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 969.977403] env[62600]: value = "task-1222869" [ 969.977403] env[62600]: _type = "Task" [ 969.977403] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.980782] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.888s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.983022] env[62600]: DEBUG oslo_concurrency.lockutils [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.240s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.983276] env[62600]: DEBUG nova.objects.instance [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lazy-loading 'resources' on Instance uuid 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.989703] env[62600]: DEBUG oslo_vmware.api [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222869, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.011293] env[62600]: INFO nova.scheduler.client.report [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Deleted allocations for instance 784e6d19-36ef-4c01-9e9f-5a083ed08608 [ 970.303256] env[62600]: DEBUG oslo_vmware.api [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222867, 'name': PowerOnVM_Task, 'duration_secs': 0.466275} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.303675] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 970.303803] env[62600]: INFO nova.compute.manager [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Took 8.24 seconds to spawn the instance on the hypervisor. [ 970.304009] env[62600]: DEBUG nova.compute.manager [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.304903] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8f65a2b-bec7-46c0-b7c6-c8ad15d79415 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.377120] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222868, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.493097] env[62600]: DEBUG oslo_vmware.api [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222869, 'name': PowerOffVM_Task, 'duration_secs': 0.328466} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.497012] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 970.497012] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 970.497012] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ff5f305d-5742-42ee-ac63-5e2cefccffc0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.523633] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f3c12395-5e0b-4e54-bf6a-4de087dea0c5 tempest-MultipleCreateTestJSON-1371801611 tempest-MultipleCreateTestJSON-1371801611-project-member] Lock "784e6d19-36ef-4c01-9e9f-5a083ed08608" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.432s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.565482] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 970.565719] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 970.565911] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleting the datastore file [datastore1] 4d12c133-84d7-4ce9-83db-e4b70b8947f3 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 970.566200] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-12fc38ae-ee0e-45ff-8738-d7ffeb227014 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.572434] env[62600]: DEBUG oslo_vmware.api [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 970.572434] env[62600]: value = "task-1222871" [ 970.572434] env[62600]: _type = "Task" [ 970.572434] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.584018] env[62600]: DEBUG oslo_vmware.api [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222871, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.675036] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fc51da2-87c0-4858-b61b-79004fd5234a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.681826] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d49f1d4-9315-4bf9-bce7-bd84d95b1b34 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.713516] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ca9eec-087e-44c2-860a-fa4c166c7ec3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.721435] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Acquiring lock "310a9093-4066-4bc3-af94-5d606860dbad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.721803] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Lock "310a9093-4066-4bc3-af94-5d606860dbad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.727329] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3830a84-3913-4672-a15b-430b2cc05d7a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.743985] env[62600]: DEBUG nova.compute.provider_tree [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.824157] env[62600]: INFO nova.compute.manager [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Took 13.67 seconds to build instance. [ 970.878363] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222868, 'name': ReconfigVM_Task, 'duration_secs': 0.864092} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.878363] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 248a5371-6ff7-4da7-ae97-d638bea0123a/e7bd1d77-bfff-4684-9545-f6a1c69efe58-rescue.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 970.878363] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37822ca9-1973-4c01-8449-34f3fd6e7119 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.906588] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d56f9bb3-2736-4e86-96ad-1270ad428f16 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.922868] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 970.922868] env[62600]: value = "task-1222872" [ 970.922868] env[62600]: _type = "Task" [ 970.922868] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.931122] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222872, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.084264] env[62600]: DEBUG oslo_vmware.api [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222871, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170343} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.084534] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 971.084801] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 971.084918] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 971.085099] env[62600]: INFO nova.compute.manager [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Took 1.12 seconds to destroy the instance on the hypervisor. [ 971.085472] env[62600]: DEBUG oslo.service.loopingcall [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 971.085795] env[62600]: DEBUG nova.compute.manager [-] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 971.085795] env[62600]: DEBUG nova.network.neutron [-] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 971.229491] env[62600]: DEBUG nova.compute.manager [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 971.249203] env[62600]: DEBUG nova.scheduler.client.report [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 971.326456] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cf19cbaf-7c07-458a-8fcb-2a84e1848f19 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "02c6dac9-535b-436b-bbee-c8031cb5d7f5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.193s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.438811] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222872, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.482354] env[62600]: DEBUG nova.compute.manager [req-a73a0b5e-94d7-4a7c-9821-6bcc2494fb43 req-9004cb2d-7152-4314-9c65-db897a5038d0 service nova] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Received event network-vif-deleted-a3a87739-c27f-4a13-a4f3-90ac701486e8 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 971.482872] env[62600]: INFO nova.compute.manager [req-a73a0b5e-94d7-4a7c-9821-6bcc2494fb43 req-9004cb2d-7152-4314-9c65-db897a5038d0 service nova] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Neutron deleted interface a3a87739-c27f-4a13-a4f3-90ac701486e8; detaching it from the instance and deleting it from the info cache [ 971.483529] env[62600]: DEBUG nova.network.neutron [req-a73a0b5e-94d7-4a7c-9821-6bcc2494fb43 req-9004cb2d-7152-4314-9c65-db897a5038d0 service nova] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.753444] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.754264] env[62600]: DEBUG oslo_concurrency.lockutils [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.771s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.756435] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.003s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.763312] env[62600]: INFO nova.compute.claims [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 971.777127] env[62600]: INFO nova.scheduler.client.report [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Deleted allocations for instance 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57 [ 971.934256] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222872, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.941719] env[62600]: DEBUG nova.network.neutron [-] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.986692] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8cf3b5b2-24af-4618-8b71-0f6f1d3bd68c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.996323] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a681ee8c-542d-4a0d-aee5-65b725422e71 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.024341] env[62600]: DEBUG nova.compute.manager [req-a73a0b5e-94d7-4a7c-9821-6bcc2494fb43 req-9004cb2d-7152-4314-9c65-db897a5038d0 service nova] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Detach interface failed, port_id=a3a87739-c27f-4a13-a4f3-90ac701486e8, reason: Instance 4d12c133-84d7-4ce9-83db-e4b70b8947f3 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 972.287972] env[62600]: DEBUG oslo_concurrency.lockutils [None req-147b9cff-23ab-4a82-b329-3a1df2b19fad tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "234dc7f8-e2ef-4f20-b25b-dcd61ce20c57" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.206s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.434935] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222872, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.444078] env[62600]: INFO nova.compute.manager [-] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Took 1.36 seconds to deallocate network for instance. [ 972.911178] env[62600]: DEBUG nova.compute.manager [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Stashing vm_state: active {{(pid=62600) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 972.937123] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222872, 'name': ReconfigVM_Task, 'duration_secs': 1.607158} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.940174] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 972.940914] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b6e2fbfe-c599-4168-8b1f-a4f1f376f3ba {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.947032] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 972.947032] env[62600]: value = "task-1222873" [ 972.947032] env[62600]: _type = "Task" [ 972.947032] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.953804] env[62600]: DEBUG oslo_concurrency.lockutils [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.954648] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff6031de-ac43-4074-895f-8f5284c9107d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.962623] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222873, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.965524] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-632de9d6-b72f-4b1b-9b42-6387a3245405 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.999032] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a7bb4d3-9bdf-4fd5-a5f3-50f4a6695d0f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.007873] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deaedd70-75eb-4240-9432-c4b6cdaa1d02 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.022946] env[62600]: DEBUG nova.compute.provider_tree [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.242297] env[62600]: DEBUG oslo_concurrency.lockutils [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Acquiring lock "e4bc0326-66d8-45e2-bcd8-11c4fab16e34" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.242534] env[62600]: DEBUG oslo_concurrency.lockutils [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Lock "e4bc0326-66d8-45e2-bcd8-11c4fab16e34" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.429389] env[62600]: DEBUG oslo_concurrency.lockutils [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.458189] env[62600]: DEBUG oslo_vmware.api [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222873, 'name': PowerOnVM_Task, 'duration_secs': 0.491894} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.458525] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 973.463451] env[62600]: DEBUG nova.compute.manager [None req-d160d6bf-0625-4776-8bc7-96b47b70fad9 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 973.464568] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e751f28-0c98-48ae-b3b5-d8adea15c1b4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.527539] env[62600]: DEBUG nova.scheduler.client.report [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 973.617131] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "aff91190-e19b-4f1c-8b47-9539360a4596" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.617399] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "aff91190-e19b-4f1c-8b47-9539360a4596" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.745301] env[62600]: DEBUG nova.compute.manager [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 974.033399] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.277s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.033992] env[62600]: DEBUG nova.compute.manager [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 974.036406] env[62600]: DEBUG oslo_concurrency.lockutils [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.083s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.037022] env[62600]: DEBUG nova.objects.instance [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lazy-loading 'resources' on Instance uuid 4d12c133-84d7-4ce9-83db-e4b70b8947f3 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.119759] env[62600]: DEBUG nova.compute.manager [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 974.266470] env[62600]: DEBUG oslo_concurrency.lockutils [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.308669] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d8665aeb-8448-411b-aa46-f4718a47439a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "interface-88cf521d-f4a3-493a-894d-c7f799dc0443-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.308938] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d8665aeb-8448-411b-aa46-f4718a47439a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "interface-88cf521d-f4a3-493a-894d-c7f799dc0443-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.309299] env[62600]: DEBUG nova.objects.instance [None req-d8665aeb-8448-411b-aa46-f4718a47439a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lazy-loading 'flavor' on Instance uuid 88cf521d-f4a3-493a-894d-c7f799dc0443 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.465756] env[62600]: INFO nova.compute.manager [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Unrescuing [ 974.466107] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.466323] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquired lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.466519] env[62600]: DEBUG nova.network.neutron [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 974.539566] env[62600]: DEBUG nova.compute.utils [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 974.544212] env[62600]: DEBUG nova.compute.manager [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 974.544404] env[62600]: DEBUG nova.network.neutron [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 974.593065] env[62600]: DEBUG nova.policy [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ff6cce064c944335b363e92d768f3c93', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c34b87fbde1443ea21fe89b7ab86dac', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 974.637462] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.702319] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a45a17-ecf0-43ff-a08f-5df5da5b899c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.710231] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9549d08-0d1a-430c-b966-a2e84747f020 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.742050] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a252739a-0d69-40fc-acc6-fdaf3bb85d8b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.749544] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca7e9270-509c-4b1a-895a-deb61055016c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.762391] env[62600]: DEBUG nova.compute.provider_tree [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.814927] env[62600]: DEBUG nova.objects.instance [None req-d8665aeb-8448-411b-aa46-f4718a47439a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lazy-loading 'pci_requests' on Instance uuid 88cf521d-f4a3-493a-894d-c7f799dc0443 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.858794] env[62600]: DEBUG nova.network.neutron [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Successfully created port: 038c3f9f-2b55-451c-abfe-33f1ca92935d {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 975.045377] env[62600]: DEBUG nova.compute.manager [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 975.265677] env[62600]: DEBUG nova.scheduler.client.report [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 975.316426] env[62600]: DEBUG nova.objects.base [None req-d8665aeb-8448-411b-aa46-f4718a47439a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Object Instance<88cf521d-f4a3-493a-894d-c7f799dc0443> lazy-loaded attributes: flavor,pci_requests {{(pid=62600) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 975.316664] env[62600]: DEBUG nova.network.neutron [None req-d8665aeb-8448-411b-aa46-f4718a47439a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 975.403753] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d8665aeb-8448-411b-aa46-f4718a47439a tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "interface-88cf521d-f4a3-493a-894d-c7f799dc0443-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.095s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.555156] env[62600]: DEBUG nova.network.neutron [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Updating instance_info_cache with network_info: [{"id": "32f5d56d-a029-4469-ad3c-77a4ca51bc02", "address": "fa:16:3e:da:8c:98", "network": {"id": "036dc114-6ea0-46cb-a958-4a6137e7e85f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-728627542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e1772c9a6c4671ab89df2e8b40cae1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32f5d56d-a0", "ovs_interfaceid": "32f5d56d-a029-4469-ad3c-77a4ca51bc02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.770796] env[62600]: DEBUG oslo_concurrency.lockutils [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.734s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.773021] env[62600]: DEBUG oslo_concurrency.lockutils [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.344s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.796496] env[62600]: INFO nova.scheduler.client.report [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleted allocations for instance 4d12c133-84d7-4ce9-83db-e4b70b8947f3 [ 976.058055] env[62600]: DEBUG nova.compute.manager [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 976.060824] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Releasing lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.060824] env[62600]: DEBUG nova.objects.instance [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lazy-loading 'flavor' on Instance uuid 248a5371-6ff7-4da7-ae97-d638bea0123a {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.083304] env[62600]: DEBUG nova.virt.hardware [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 976.083583] env[62600]: DEBUG nova.virt.hardware [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 976.083751] env[62600]: DEBUG nova.virt.hardware [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 976.083975] env[62600]: DEBUG nova.virt.hardware [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 976.084148] env[62600]: DEBUG nova.virt.hardware [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 976.084302] env[62600]: DEBUG nova.virt.hardware [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 976.084516] env[62600]: DEBUG nova.virt.hardware [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 976.084684] env[62600]: DEBUG nova.virt.hardware [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 976.084934] env[62600]: DEBUG nova.virt.hardware [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 976.085135] env[62600]: DEBUG nova.virt.hardware [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 976.085343] env[62600]: DEBUG nova.virt.hardware [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 976.086513] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc4df35-de51-4fc3-bab2-b3cc0e51a700 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.094874] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54fd66c9-e4a9-47e7-9ccb-698bc7c3069c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.277874] env[62600]: INFO nova.compute.claims [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 976.303663] env[62600]: DEBUG oslo_concurrency.lockutils [None req-34d93787-7aaa-48ec-ae64-8d3ef38de8bb tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "4d12c133-84d7-4ce9-83db-e4b70b8947f3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.349s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.567196] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c862f6-7349-4827-b83f-0e4df926c1b3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.593195] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 976.593481] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ce5c1985-e88b-4bf2-a3bf-cf00c870cf0c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.599745] env[62600]: DEBUG oslo_vmware.api [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 976.599745] env[62600]: value = "task-1222874" [ 976.599745] env[62600]: _type = "Task" [ 976.599745] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.607377] env[62600]: DEBUG oslo_vmware.api [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222874, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.784768] env[62600]: INFO nova.compute.resource_tracker [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Updating resource usage from migration 8d35c436-258f-4615-a3a2-eb29f7f59091 [ 976.840797] env[62600]: DEBUG nova.compute.manager [req-7de93a61-5f37-49e8-99e6-96732dd18e24 req-2f49ed0b-8b52-4d54-8118-094c295366f3 service nova] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Received event network-vif-plugged-038c3f9f-2b55-451c-abfe-33f1ca92935d {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.841104] env[62600]: DEBUG oslo_concurrency.lockutils [req-7de93a61-5f37-49e8-99e6-96732dd18e24 req-2f49ed0b-8b52-4d54-8118-094c295366f3 service nova] Acquiring lock "310a9093-4066-4bc3-af94-5d606860dbad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.841283] env[62600]: DEBUG oslo_concurrency.lockutils [req-7de93a61-5f37-49e8-99e6-96732dd18e24 req-2f49ed0b-8b52-4d54-8118-094c295366f3 service nova] Lock "310a9093-4066-4bc3-af94-5d606860dbad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.841488] env[62600]: DEBUG oslo_concurrency.lockutils [req-7de93a61-5f37-49e8-99e6-96732dd18e24 req-2f49ed0b-8b52-4d54-8118-094c295366f3 service nova] Lock "310a9093-4066-4bc3-af94-5d606860dbad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.841654] env[62600]: DEBUG nova.compute.manager [req-7de93a61-5f37-49e8-99e6-96732dd18e24 req-2f49ed0b-8b52-4d54-8118-094c295366f3 service nova] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] No waiting events found dispatching network-vif-plugged-038c3f9f-2b55-451c-abfe-33f1ca92935d {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 976.841846] env[62600]: WARNING nova.compute.manager [req-7de93a61-5f37-49e8-99e6-96732dd18e24 req-2f49ed0b-8b52-4d54-8118-094c295366f3 service nova] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Received unexpected event network-vif-plugged-038c3f9f-2b55-451c-abfe-33f1ca92935d for instance with vm_state building and task_state spawning. [ 976.873235] env[62600]: DEBUG nova.network.neutron [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Successfully updated port: 038c3f9f-2b55-451c-abfe-33f1ca92935d {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 976.992028] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15470cd2-44e4-48c2-9818-cc5283452c6c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.999302] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15761946-6c2b-4a30-9d97-6eb4b7bbf2d9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.030431] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c6e057-cc2f-4726-ab70-b5f63d06db75 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.037902] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0349c11e-0eaf-4ab5-862d-5f90a256df0b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.051225] env[62600]: DEBUG nova.compute.provider_tree [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.109239] env[62600]: DEBUG oslo_vmware.api [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222874, 'name': PowerOffVM_Task, 'duration_secs': 0.220037} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.109501] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 977.115391] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Reconfiguring VM instance instance-00000050 to detach disk 2002 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 977.115836] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3919b031-7aa4-42b4-b786-e85f4992b799 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.135034] env[62600]: DEBUG oslo_vmware.api [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 977.135034] env[62600]: value = "task-1222875" [ 977.135034] env[62600]: _type = "Task" [ 977.135034] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.142908] env[62600]: DEBUG oslo_vmware.api [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222875, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.380767] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Acquiring lock "refresh_cache-310a9093-4066-4bc3-af94-5d606860dbad" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.380892] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Acquired lock "refresh_cache-310a9093-4066-4bc3-af94-5d606860dbad" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.381067] env[62600]: DEBUG nova.network.neutron [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 977.382446] env[62600]: DEBUG oslo_concurrency.lockutils [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "interface-88cf521d-f4a3-493a-894d-c7f799dc0443-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.386214] env[62600]: DEBUG oslo_concurrency.lockutils [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "interface-88cf521d-f4a3-493a-894d-c7f799dc0443-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.386214] env[62600]: DEBUG nova.objects.instance [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lazy-loading 'flavor' on Instance uuid 88cf521d-f4a3-493a-894d-c7f799dc0443 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.554445] env[62600]: DEBUG nova.scheduler.client.report [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 977.644981] env[62600]: DEBUG oslo_vmware.api [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222875, 'name': ReconfigVM_Task, 'duration_secs': 0.234469} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.645290] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Reconfigured VM instance instance-00000050 to detach disk 2002 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 977.645532] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 977.645835] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-280543b1-2c3c-4815-bb72-09b89a6981ef {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.652233] env[62600]: DEBUG oslo_vmware.api [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 977.652233] env[62600]: value = "task-1222876" [ 977.652233] env[62600]: _type = "Task" [ 977.652233] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.661517] env[62600]: DEBUG oslo_vmware.api [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222876, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.725208] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "5d7c2a17-9dee-407e-a7ee-6bb88f56905f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.725520] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "5d7c2a17-9dee-407e-a7ee-6bb88f56905f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.941032] env[62600]: DEBUG nova.network.neutron [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 978.059090] env[62600]: DEBUG oslo_concurrency.lockutils [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.286s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.059342] env[62600]: INFO nova.compute.manager [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Migrating [ 978.068582] env[62600]: DEBUG oslo_concurrency.lockutils [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.802s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.070101] env[62600]: INFO nova.compute.claims [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 978.080957] env[62600]: DEBUG nova.objects.instance [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lazy-loading 'pci_requests' on Instance uuid 88cf521d-f4a3-493a-894d-c7f799dc0443 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.148208] env[62600]: DEBUG nova.network.neutron [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Updating instance_info_cache with network_info: [{"id": "038c3f9f-2b55-451c-abfe-33f1ca92935d", "address": "fa:16:3e:f8:83:ef", "network": {"id": "4790417e-470a-4d86-81c5-7ec8dff98008", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-227567693-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c34b87fbde1443ea21fe89b7ab86dac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap038c3f9f-2b", "ovs_interfaceid": "038c3f9f-2b55-451c-abfe-33f1ca92935d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.161828] env[62600]: DEBUG oslo_vmware.api [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222876, 'name': PowerOnVM_Task, 'duration_secs': 0.397798} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.162105] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 978.162344] env[62600]: DEBUG nova.compute.manager [None req-3fa4e8a5-9f32-406d-965e-955857cbc332 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 978.163114] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a3e8f2-b8f2-42c8-85ed-403a56991b69 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.227633] env[62600]: DEBUG nova.compute.manager [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 978.583530] env[62600]: DEBUG oslo_concurrency.lockutils [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "refresh_cache-02c6dac9-535b-436b-bbee-c8031cb5d7f5" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.583707] env[62600]: DEBUG oslo_concurrency.lockutils [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "refresh_cache-02c6dac9-535b-436b-bbee-c8031cb5d7f5" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.583971] env[62600]: DEBUG nova.network.neutron [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 978.585157] env[62600]: DEBUG nova.objects.base [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Object Instance<88cf521d-f4a3-493a-894d-c7f799dc0443> lazy-loaded attributes: flavor,pci_requests {{(pid=62600) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 978.585349] env[62600]: DEBUG nova.network.neutron [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 978.621650] env[62600]: DEBUG nova.policy [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fbd1b2fb34d841359ada8fc44bec2986', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b028450f2da445fb83e37adfc86bba68', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 978.651080] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Releasing lock "refresh_cache-310a9093-4066-4bc3-af94-5d606860dbad" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.651567] env[62600]: DEBUG nova.compute.manager [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Instance network_info: |[{"id": "038c3f9f-2b55-451c-abfe-33f1ca92935d", "address": "fa:16:3e:f8:83:ef", "network": {"id": "4790417e-470a-4d86-81c5-7ec8dff98008", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-227567693-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c34b87fbde1443ea21fe89b7ab86dac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap038c3f9f-2b", "ovs_interfaceid": "038c3f9f-2b55-451c-abfe-33f1ca92935d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 978.652039] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f8:83:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c9bc2632-36f9-4912-8782-8bbb789f909d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '038c3f9f-2b55-451c-abfe-33f1ca92935d', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 978.659868] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Creating folder: Project (4c34b87fbde1443ea21fe89b7ab86dac). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 978.660444] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-656c12fe-ed0b-4079-8143-1faeedc91220 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.670999] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Created folder: Project (4c34b87fbde1443ea21fe89b7ab86dac) in parent group-v264198. [ 978.671220] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Creating folder: Instances. Parent ref: group-v264353. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 978.673461] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1a383123-35f8-438d-a107-6d9322c7803b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.682791] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Created folder: Instances in parent group-v264353. [ 978.683031] env[62600]: DEBUG oslo.service.loopingcall [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 978.683245] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 978.683452] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b9913269-6e9d-4ead-9908-5aa92dcc2ecd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.702253] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 978.702253] env[62600]: value = "task-1222879" [ 978.702253] env[62600]: _type = "Task" [ 978.702253] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.709785] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222879, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.754602] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.955261] env[62600]: DEBUG nova.compute.manager [req-f03c5737-3b0d-497a-8ed7-73d5cef46686 req-732b1124-7d8a-4032-befa-c0753ca36209 service nova] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Received event network-changed-038c3f9f-2b55-451c-abfe-33f1ca92935d {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.955472] env[62600]: DEBUG nova.compute.manager [req-f03c5737-3b0d-497a-8ed7-73d5cef46686 req-732b1124-7d8a-4032-befa-c0753ca36209 service nova] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Refreshing instance network info cache due to event network-changed-038c3f9f-2b55-451c-abfe-33f1ca92935d. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 978.955703] env[62600]: DEBUG oslo_concurrency.lockutils [req-f03c5737-3b0d-497a-8ed7-73d5cef46686 req-732b1124-7d8a-4032-befa-c0753ca36209 service nova] Acquiring lock "refresh_cache-310a9093-4066-4bc3-af94-5d606860dbad" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.955894] env[62600]: DEBUG oslo_concurrency.lockutils [req-f03c5737-3b0d-497a-8ed7-73d5cef46686 req-732b1124-7d8a-4032-befa-c0753ca36209 service nova] Acquired lock "refresh_cache-310a9093-4066-4bc3-af94-5d606860dbad" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.958815] env[62600]: DEBUG nova.network.neutron [req-f03c5737-3b0d-497a-8ed7-73d5cef46686 req-732b1124-7d8a-4032-befa-c0753ca36209 service nova] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Refreshing network info cache for port 038c3f9f-2b55-451c-abfe-33f1ca92935d {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 978.981136] env[62600]: DEBUG nova.network.neutron [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Successfully created port: 24f6c3e5-a847-44bd-ba62-c4d66a2489d8 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 979.215371] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222879, 'name': CreateVM_Task, 'duration_secs': 0.365113} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.218097] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 979.219855] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.220088] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.220438] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 979.220707] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffa5e51c-72f2-420d-b380-e4ee8cfc8ba8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.228332] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Waiting for the task: (returnval){ [ 979.228332] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d74910-4b15-160b-8221-47acad94c95b" [ 979.228332] env[62600]: _type = "Task" [ 979.228332] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.244614] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d74910-4b15-160b-8221-47acad94c95b, 'name': SearchDatastore_Task, 'duration_secs': 0.009598} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.244909] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.245212] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 979.245479] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.245637] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.245823] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 979.246097] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-98a66d07-776a-4dc3-9dd5-74b3f03a611c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.253798] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 979.254078] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 979.254792] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee662340-af17-43e9-9e9b-7ab4cbd0bec4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.277196] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Waiting for the task: (returnval){ [ 979.277196] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52cfb11c-aa90-8f1e-185a-d7575a0a5d1c" [ 979.277196] env[62600]: _type = "Task" [ 979.277196] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.277196] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04af96b3-b8e6-4bdf-8fd2-1a380d4a0609 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.292676] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748c47dc-89cf-41cb-968d-1d65d0d895fb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.295700] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52cfb11c-aa90-8f1e-185a-d7575a0a5d1c, 'name': SearchDatastore_Task, 'duration_secs': 0.009534} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.300022] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6b643ab-7097-4b94-bd11-78bc444f6b19 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.323588] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 979.328263] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 979.328263] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Starting heal instance info cache {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 979.328733] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73780785-fea7-411e-9b9f-6000216e7dc0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.338337] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Waiting for the task: (returnval){ [ 979.338337] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b638af-2a16-e62b-9a3a-e85377a4ae83" [ 979.338337] env[62600]: _type = "Task" [ 979.338337] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.341490] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5517879f-ed0e-473d-9d07-a3a1a4b093e4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.349916] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b638af-2a16-e62b-9a3a-e85377a4ae83, 'name': SearchDatastore_Task, 'duration_secs': 0.011431} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.350647] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.350946] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 310a9093-4066-4bc3-af94-5d606860dbad/310a9093-4066-4bc3-af94-5d606860dbad.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 979.351236] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-836a9a9a-330b-470e-b838-9dd8e23297f8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.361440] env[62600]: DEBUG nova.compute.provider_tree [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 979.368806] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Waiting for the task: (returnval){ [ 979.368806] env[62600]: value = "task-1222880" [ 979.368806] env[62600]: _type = "Task" [ 979.368806] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.380813] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Task: {'id': task-1222880, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.465248] env[62600]: DEBUG nova.network.neutron [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Updating instance_info_cache with network_info: [{"id": "3753b6a6-f27d-4e1d-b801-91324a7e4e0e", "address": "fa:16:3e:fd:6e:a3", "network": {"id": "95818650-8227-4d11-9d14-007f06c07790", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1738152115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73e664d178f7484a9f4741b4d9450e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3753b6a6-f2", "ovs_interfaceid": "3753b6a6-f27d-4e1d-b801-91324a7e4e0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.603218] env[62600]: DEBUG nova.compute.manager [req-f1285233-9024-419c-9bff-81f589d53d55 req-0375f639-e5ad-4322-8fe9-b6078d01ad56 service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Received event network-changed-32f5d56d-a029-4469-ad3c-77a4ca51bc02 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 979.603218] env[62600]: DEBUG nova.compute.manager [req-f1285233-9024-419c-9bff-81f589d53d55 req-0375f639-e5ad-4322-8fe9-b6078d01ad56 service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Refreshing instance network info cache due to event network-changed-32f5d56d-a029-4469-ad3c-77a4ca51bc02. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 979.603218] env[62600]: DEBUG oslo_concurrency.lockutils [req-f1285233-9024-419c-9bff-81f589d53d55 req-0375f639-e5ad-4322-8fe9-b6078d01ad56 service nova] Acquiring lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.603218] env[62600]: DEBUG oslo_concurrency.lockutils [req-f1285233-9024-419c-9bff-81f589d53d55 req-0375f639-e5ad-4322-8fe9-b6078d01ad56 service nova] Acquired lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.603416] env[62600]: DEBUG nova.network.neutron [req-f1285233-9024-419c-9bff-81f589d53d55 req-0375f639-e5ad-4322-8fe9-b6078d01ad56 service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Refreshing network info cache for port 32f5d56d-a029-4469-ad3c-77a4ca51bc02 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 979.744557] env[62600]: DEBUG nova.network.neutron [req-f03c5737-3b0d-497a-8ed7-73d5cef46686 req-732b1124-7d8a-4032-befa-c0753ca36209 service nova] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Updated VIF entry in instance network info cache for port 038c3f9f-2b55-451c-abfe-33f1ca92935d. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 979.745101] env[62600]: DEBUG nova.network.neutron [req-f03c5737-3b0d-497a-8ed7-73d5cef46686 req-732b1124-7d8a-4032-befa-c0753ca36209 service nova] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Updating instance_info_cache with network_info: [{"id": "038c3f9f-2b55-451c-abfe-33f1ca92935d", "address": "fa:16:3e:f8:83:ef", "network": {"id": "4790417e-470a-4d86-81c5-7ec8dff98008", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-227567693-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4c34b87fbde1443ea21fe89b7ab86dac", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap038c3f9f-2b", "ovs_interfaceid": "038c3f9f-2b55-451c-abfe-33f1ca92935d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.865021] env[62600]: DEBUG nova.scheduler.client.report [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 979.879207] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Task: {'id': task-1222880, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500504} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.879539] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 310a9093-4066-4bc3-af94-5d606860dbad/310a9093-4066-4bc3-af94-5d606860dbad.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 979.879788] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 979.880755] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-928a93aa-2ba1-436b-88cb-560be0ed5d6e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.887490] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Waiting for the task: (returnval){ [ 979.887490] env[62600]: value = "task-1222881" [ 979.887490] env[62600]: _type = "Task" [ 979.887490] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.895575] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Task: {'id': task-1222881, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.968724] env[62600]: DEBUG oslo_concurrency.lockutils [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "refresh_cache-02c6dac9-535b-436b-bbee-c8031cb5d7f5" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.250130] env[62600]: DEBUG oslo_concurrency.lockutils [req-f03c5737-3b0d-497a-8ed7-73d5cef46686 req-732b1124-7d8a-4032-befa-c0753ca36209 service nova] Releasing lock "refresh_cache-310a9093-4066-4bc3-af94-5d606860dbad" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.321270] env[62600]: DEBUG nova.network.neutron [req-f1285233-9024-419c-9bff-81f589d53d55 req-0375f639-e5ad-4322-8fe9-b6078d01ad56 service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Updated VIF entry in instance network info cache for port 32f5d56d-a029-4469-ad3c-77a4ca51bc02. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 980.321655] env[62600]: DEBUG nova.network.neutron [req-f1285233-9024-419c-9bff-81f589d53d55 req-0375f639-e5ad-4322-8fe9-b6078d01ad56 service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Updating instance_info_cache with network_info: [{"id": "32f5d56d-a029-4469-ad3c-77a4ca51bc02", "address": "fa:16:3e:da:8c:98", "network": {"id": "036dc114-6ea0-46cb-a958-4a6137e7e85f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-728627542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e1772c9a6c4671ab89df2e8b40cae1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32f5d56d-a0", "ovs_interfaceid": "32f5d56d-a029-4469-ad3c-77a4ca51bc02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.370355] env[62600]: DEBUG oslo_concurrency.lockutils [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.302s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.370884] env[62600]: DEBUG nova.compute.manager [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 980.374121] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.736s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.375064] env[62600]: INFO nova.compute.claims [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 980.401559] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Task: {'id': task-1222881, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074199} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.401559] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 980.402353] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f01f6cfa-5cef-487d-bc64-be97c3a16015 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.426563] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] 310a9093-4066-4bc3-af94-5d606860dbad/310a9093-4066-4bc3-af94-5d606860dbad.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 980.427166] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3bf5cc13-35f4-4b63-872a-f94f1552bd1f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.447451] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Waiting for the task: (returnval){ [ 980.447451] env[62600]: value = "task-1222882" [ 980.447451] env[62600]: _type = "Task" [ 980.447451] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.457314] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Task: {'id': task-1222882, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.736116] env[62600]: DEBUG nova.network.neutron [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Successfully updated port: 24f6c3e5-a847-44bd-ba62-c4d66a2489d8 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 980.824486] env[62600]: DEBUG oslo_concurrency.lockutils [req-f1285233-9024-419c-9bff-81f589d53d55 req-0375f639-e5ad-4322-8fe9-b6078d01ad56 service nova] Releasing lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.879191] env[62600]: DEBUG nova.compute.utils [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 980.882564] env[62600]: DEBUG nova.compute.manager [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Not allocating networking since 'none' was specified. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 980.958863] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Task: {'id': task-1222882, 'name': ReconfigVM_Task, 'duration_secs': 0.292181} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.959243] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Reconfigured VM instance instance-0000005b to attach disk [datastore2] 310a9093-4066-4bc3-af94-5d606860dbad/310a9093-4066-4bc3-af94-5d606860dbad.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 980.959909] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-02b040cb-cad8-41ee-80df-1cb7348411a0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.967146] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Waiting for the task: (returnval){ [ 980.967146] env[62600]: value = "task-1222883" [ 980.967146] env[62600]: _type = "Task" [ 980.967146] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.978687] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Task: {'id': task-1222883, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.010947] env[62600]: DEBUG nova.compute.manager [req-180aa60b-387a-4c60-99b9-2ef2ed28b4fc req-b08f9e80-ba50-49fb-87b6-730dfe143b35 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Received event network-vif-plugged-24f6c3e5-a847-44bd-ba62-c4d66a2489d8 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 981.011149] env[62600]: DEBUG oslo_concurrency.lockutils [req-180aa60b-387a-4c60-99b9-2ef2ed28b4fc req-b08f9e80-ba50-49fb-87b6-730dfe143b35 service nova] Acquiring lock "88cf521d-f4a3-493a-894d-c7f799dc0443-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.011378] env[62600]: DEBUG oslo_concurrency.lockutils [req-180aa60b-387a-4c60-99b9-2ef2ed28b4fc req-b08f9e80-ba50-49fb-87b6-730dfe143b35 service nova] Lock "88cf521d-f4a3-493a-894d-c7f799dc0443-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.011559] env[62600]: DEBUG oslo_concurrency.lockutils [req-180aa60b-387a-4c60-99b9-2ef2ed28b4fc req-b08f9e80-ba50-49fb-87b6-730dfe143b35 service nova] Lock "88cf521d-f4a3-493a-894d-c7f799dc0443-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.011733] env[62600]: DEBUG nova.compute.manager [req-180aa60b-387a-4c60-99b9-2ef2ed28b4fc req-b08f9e80-ba50-49fb-87b6-730dfe143b35 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] No waiting events found dispatching network-vif-plugged-24f6c3e5-a847-44bd-ba62-c4d66a2489d8 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 981.011899] env[62600]: WARNING nova.compute.manager [req-180aa60b-387a-4c60-99b9-2ef2ed28b4fc req-b08f9e80-ba50-49fb-87b6-730dfe143b35 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Received unexpected event network-vif-plugged-24f6c3e5-a847-44bd-ba62-c4d66a2489d8 for instance with vm_state active and task_state None. [ 981.012074] env[62600]: DEBUG nova.compute.manager [req-180aa60b-387a-4c60-99b9-2ef2ed28b4fc req-b08f9e80-ba50-49fb-87b6-730dfe143b35 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Received event network-changed-24f6c3e5-a847-44bd-ba62-c4d66a2489d8 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 981.012245] env[62600]: DEBUG nova.compute.manager [req-180aa60b-387a-4c60-99b9-2ef2ed28b4fc req-b08f9e80-ba50-49fb-87b6-730dfe143b35 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Refreshing instance network info cache due to event network-changed-24f6c3e5-a847-44bd-ba62-c4d66a2489d8. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 981.012430] env[62600]: DEBUG oslo_concurrency.lockutils [req-180aa60b-387a-4c60-99b9-2ef2ed28b4fc req-b08f9e80-ba50-49fb-87b6-730dfe143b35 service nova] Acquiring lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.012571] env[62600]: DEBUG oslo_concurrency.lockutils [req-180aa60b-387a-4c60-99b9-2ef2ed28b4fc req-b08f9e80-ba50-49fb-87b6-730dfe143b35 service nova] Acquired lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.012732] env[62600]: DEBUG nova.network.neutron [req-180aa60b-387a-4c60-99b9-2ef2ed28b4fc req-b08f9e80-ba50-49fb-87b6-730dfe143b35 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Refreshing network info cache for port 24f6c3e5-a847-44bd-ba62-c4d66a2489d8 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 981.238633] env[62600]: DEBUG oslo_concurrency.lockutils [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.383298] env[62600]: DEBUG nova.compute.manager [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 981.478166] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Task: {'id': task-1222883, 'name': Rename_Task, 'duration_secs': 0.139502} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.480544] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 981.482988] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d2aefd1-9f8f-4b6b-9817-af6701d94618 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.487425] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92f43e0f-9e7c-42f7-b2a3-8d6d87971512 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.509465] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Updating instance '02c6dac9-535b-436b-bbee-c8031cb5d7f5' progress to 0 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 981.511863] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Waiting for the task: (returnval){ [ 981.511863] env[62600]: value = "task-1222884" [ 981.511863] env[62600]: _type = "Task" [ 981.511863] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.525383] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Task: {'id': task-1222884, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.596584] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a239f7b4-d270-436a-a3a7-4f8d2c38ace8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.604781] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-991c6271-6299-47c2-bb4e-3896dc67772e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.636292] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b793b963-1044-4e8d-9a76-eacf1ddd7a0e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.644112] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73c81dd1-ea73-4acd-9b7d-2a2fafca702e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.657649] env[62600]: DEBUG nova.compute.provider_tree [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 981.688160] env[62600]: DEBUG nova.compute.manager [req-d2edac27-24d2-4536-8c33-5588656d7498 req-7a75ad6e-a241-4764-903a-8d5838911dea service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Received event network-changed-32f5d56d-a029-4469-ad3c-77a4ca51bc02 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 981.688362] env[62600]: DEBUG nova.compute.manager [req-d2edac27-24d2-4536-8c33-5588656d7498 req-7a75ad6e-a241-4764-903a-8d5838911dea service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Refreshing instance network info cache due to event network-changed-32f5d56d-a029-4469-ad3c-77a4ca51bc02. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 981.688632] env[62600]: DEBUG oslo_concurrency.lockutils [req-d2edac27-24d2-4536-8c33-5588656d7498 req-7a75ad6e-a241-4764-903a-8d5838911dea service nova] Acquiring lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.688723] env[62600]: DEBUG oslo_concurrency.lockutils [req-d2edac27-24d2-4536-8c33-5588656d7498 req-7a75ad6e-a241-4764-903a-8d5838911dea service nova] Acquired lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.688916] env[62600]: DEBUG nova.network.neutron [req-d2edac27-24d2-4536-8c33-5588656d7498 req-7a75ad6e-a241-4764-903a-8d5838911dea service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Refreshing network info cache for port 32f5d56d-a029-4469-ad3c-77a4ca51bc02 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 981.787517] env[62600]: DEBUG nova.network.neutron [req-180aa60b-387a-4c60-99b9-2ef2ed28b4fc req-b08f9e80-ba50-49fb-87b6-730dfe143b35 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Added VIF to instance network info cache for port 24f6c3e5-a847-44bd-ba62-c4d66a2489d8. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3489}} [ 981.787940] env[62600]: DEBUG nova.network.neutron [req-180aa60b-387a-4c60-99b9-2ef2ed28b4fc req-b08f9e80-ba50-49fb-87b6-730dfe143b35 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Updating instance_info_cache with network_info: [{"id": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "address": "fa:16:3e:67:24:01", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8aca44-97", "ovs_interfaceid": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "24f6c3e5-a847-44bd-ba62-c4d66a2489d8", "address": "fa:16:3e:e1:5b:18", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24f6c3e5-a8", "ovs_interfaceid": "24f6c3e5-a847-44bd-ba62-c4d66a2489d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.870784] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "refresh_cache-b967fb11-e70e-4e17-b769-38da581bd83b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.871104] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquired lock "refresh_cache-b967fb11-e70e-4e17-b769-38da581bd83b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.871104] env[62600]: DEBUG nova.network.neutron [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Forcefully refreshing network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 982.014841] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 982.015250] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-18b68dac-0a4e-4a20-a9c3-d7093d9620d8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.026249] env[62600]: DEBUG oslo_vmware.api [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Task: {'id': task-1222884, 'name': PowerOnVM_Task, 'duration_secs': 0.487495} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.028063] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 982.028063] env[62600]: INFO nova.compute.manager [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Took 5.97 seconds to spawn the instance on the hypervisor. [ 982.028063] env[62600]: DEBUG nova.compute.manager [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 982.028285] env[62600]: DEBUG oslo_vmware.api [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 982.028285] env[62600]: value = "task-1222885" [ 982.028285] env[62600]: _type = "Task" [ 982.028285] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.028891] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54939a5e-0c35-441d-94ca-3585eafbde95 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.040686] env[62600]: DEBUG oslo_vmware.api [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222885, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.166277] env[62600]: DEBUG nova.scheduler.client.report [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 982.290589] env[62600]: DEBUG oslo_concurrency.lockutils [req-180aa60b-387a-4c60-99b9-2ef2ed28b4fc req-b08f9e80-ba50-49fb-87b6-730dfe143b35 service nova] Releasing lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.291014] env[62600]: DEBUG oslo_concurrency.lockutils [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.291226] env[62600]: DEBUG nova.network.neutron [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 982.393806] env[62600]: DEBUG nova.compute.manager [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 982.406376] env[62600]: DEBUG nova.network.neutron [req-d2edac27-24d2-4536-8c33-5588656d7498 req-7a75ad6e-a241-4764-903a-8d5838911dea service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Updated VIF entry in instance network info cache for port 32f5d56d-a029-4469-ad3c-77a4ca51bc02. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 982.407605] env[62600]: DEBUG nova.network.neutron [req-d2edac27-24d2-4536-8c33-5588656d7498 req-7a75ad6e-a241-4764-903a-8d5838911dea service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Updating instance_info_cache with network_info: [{"id": "32f5d56d-a029-4469-ad3c-77a4ca51bc02", "address": "fa:16:3e:da:8c:98", "network": {"id": "036dc114-6ea0-46cb-a958-4a6137e7e85f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-728627542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e1772c9a6c4671ab89df2e8b40cae1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap32f5d56d-a0", "ovs_interfaceid": "32f5d56d-a029-4469-ad3c-77a4ca51bc02", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.424881] env[62600]: DEBUG nova.virt.hardware [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 982.425154] env[62600]: DEBUG nova.virt.hardware [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 982.425328] env[62600]: DEBUG nova.virt.hardware [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 982.425520] env[62600]: DEBUG nova.virt.hardware [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 982.425672] env[62600]: DEBUG nova.virt.hardware [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 982.425826] env[62600]: DEBUG nova.virt.hardware [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 982.426046] env[62600]: DEBUG nova.virt.hardware [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 982.426219] env[62600]: DEBUG nova.virt.hardware [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 982.426390] env[62600]: DEBUG nova.virt.hardware [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 982.426554] env[62600]: DEBUG nova.virt.hardware [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 982.426730] env[62600]: DEBUG nova.virt.hardware [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 982.427864] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d7a533-35c1-4566-adaa-9cc3e0ac1a37 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.435559] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c612f50-e405-4bda-ba08-e54cacba79b3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.448877] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Instance VIF info [] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 982.454315] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Creating folder: Project (7555688a3e3a4535a55159278029cb63). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 982.454571] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-800eebf7-e772-4287-9fff-4d451241a5de {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.463784] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Created folder: Project (7555688a3e3a4535a55159278029cb63) in parent group-v264198. [ 982.463966] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Creating folder: Instances. Parent ref: group-v264356. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 982.464268] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4b7b1ea8-8153-4133-9f2d-28cc5c8d8022 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.472921] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Created folder: Instances in parent group-v264356. [ 982.473162] env[62600]: DEBUG oslo.service.loopingcall [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 982.473350] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 982.473531] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc1b7751-6963-40df-bc80-02f672876943 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.488717] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 982.488717] env[62600]: value = "task-1222888" [ 982.488717] env[62600]: _type = "Task" [ 982.488717] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.497039] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222888, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.540700] env[62600]: DEBUG oslo_vmware.api [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222885, 'name': PowerOffVM_Task, 'duration_secs': 0.205554} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.540990] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 982.541207] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Updating instance '02c6dac9-535b-436b-bbee-c8031cb5d7f5' progress to 17 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 982.550645] env[62600]: INFO nova.compute.manager [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Took 10.81 seconds to build instance. [ 982.671600] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.298s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.672190] env[62600]: DEBUG nova.compute.manager [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 982.674781] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.920s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.676537] env[62600]: INFO nova.compute.claims [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 982.831492] env[62600]: WARNING nova.network.neutron [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] 8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb already exists in list: networks containing: ['8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb']. ignoring it [ 982.831727] env[62600]: WARNING nova.network.neutron [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] 8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb already exists in list: networks containing: ['8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb']. ignoring it [ 982.831900] env[62600]: WARNING nova.network.neutron [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] 24f6c3e5-a847-44bd-ba62-c4d66a2489d8 already exists in list: port_ids containing: ['24f6c3e5-a847-44bd-ba62-c4d66a2489d8']. ignoring it [ 982.909651] env[62600]: DEBUG oslo_concurrency.lockutils [req-d2edac27-24d2-4536-8c33-5588656d7498 req-7a75ad6e-a241-4764-903a-8d5838911dea service nova] Releasing lock "refresh_cache-248a5371-6ff7-4da7-ae97-d638bea0123a" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.999056] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222888, 'name': CreateVM_Task, 'duration_secs': 0.254593} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.999056] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 982.999056] env[62600]: DEBUG oslo_concurrency.lockutils [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.999056] env[62600]: DEBUG oslo_concurrency.lockutils [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.999440] env[62600]: DEBUG oslo_concurrency.lockutils [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 982.999490] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b440bb5-c7a1-45d4-8f04-a9d874539fba {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.003851] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Waiting for the task: (returnval){ [ 983.003851] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c15701-eddd-9973-c709-8d44e5ec32c2" [ 983.003851] env[62600]: _type = "Task" [ 983.003851] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.012741] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c15701-eddd-9973-c709-8d44e5ec32c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.050616] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 983.050909] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 983.055024] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 983.055024] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 983.055024] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 983.055024] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 983.055024] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 983.055024] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 983.055024] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 983.055024] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 983.055024] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 983.059863] env[62600]: DEBUG oslo_concurrency.lockutils [None req-54b89611-44be-4b2a-a5b3-ea2c785365ab tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Lock "310a9093-4066-4bc3-af94-5d606860dbad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.338s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.060167] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2105bbc8-5be6-44ac-bd6f-ddf99658d493 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.087876] env[62600]: DEBUG oslo_vmware.api [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 983.087876] env[62600]: value = "task-1222889" [ 983.087876] env[62600]: _type = "Task" [ 983.087876] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.099054] env[62600]: DEBUG oslo_vmware.api [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222889, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.181464] env[62600]: DEBUG nova.compute.utils [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 983.186051] env[62600]: DEBUG nova.compute.manager [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 983.186051] env[62600]: DEBUG nova.network.neutron [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 983.228472] env[62600]: DEBUG nova.policy [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34319d68370c48ef93fa24a9d9132407', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bab74e3ae78248909dbcd483abbb8da2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 983.257237] env[62600]: DEBUG nova.network.neutron [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Updating instance_info_cache with network_info: [{"id": "d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049", "address": "fa:16:3e:4f:fd:77", "network": {"id": "036dc114-6ea0-46cb-a958-4a6137e7e85f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-728627542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e1772c9a6c4671ab89df2e8b40cae1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd6e6a2fb-7b", "ovs_interfaceid": "d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.311252] env[62600]: DEBUG nova.network.neutron [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Updating instance_info_cache with network_info: [{"id": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "address": "fa:16:3e:67:24:01", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8aca44-97", "ovs_interfaceid": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "24f6c3e5-a847-44bd-ba62-c4d66a2489d8", "address": "fa:16:3e:e1:5b:18", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24f6c3e5-a8", "ovs_interfaceid": "24f6c3e5-a847-44bd-ba62-c4d66a2489d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.476036] env[62600]: DEBUG nova.network.neutron [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Successfully created port: 7af8f843-5635-4c50-bc15-f8356b10e6f7 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 983.514647] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c15701-eddd-9973-c709-8d44e5ec32c2, 'name': SearchDatastore_Task, 'duration_secs': 0.009613} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.514966] env[62600]: DEBUG oslo_concurrency.lockutils [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.515589] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 983.515589] env[62600]: DEBUG oslo_concurrency.lockutils [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.515804] env[62600]: DEBUG oslo_concurrency.lockutils [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.515850] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 983.516110] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-98709f9e-1f98-4f19-97d7-659220f04e41 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.524359] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 983.524645] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 983.525279] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fa8e13d-2453-470c-a6ab-d9866ed2aa02 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.531033] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Waiting for the task: (returnval){ [ 983.531033] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52a7b197-b799-ea74-1127-e5e299c0887b" [ 983.531033] env[62600]: _type = "Task" [ 983.531033] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.539908] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52a7b197-b799-ea74-1127-e5e299c0887b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.599912] env[62600]: DEBUG oslo_vmware.api [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222889, 'name': ReconfigVM_Task, 'duration_secs': 0.165601} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.600313] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Updating instance '02c6dac9-535b-436b-bbee-c8031cb5d7f5' progress to 33 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 983.685786] env[62600]: DEBUG nova.compute.manager [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 983.759541] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Releasing lock "refresh_cache-b967fb11-e70e-4e17-b769-38da581bd83b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.759767] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Updated the network info_cache for instance {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 983.760408] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 983.760521] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 983.760607] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 983.760760] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 983.760911] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 983.761086] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 983.761514] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62600) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 983.761514] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 983.814469] env[62600]: DEBUG oslo_concurrency.lockutils [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.815118] env[62600]: DEBUG oslo_concurrency.lockutils [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.815527] env[62600]: DEBUG oslo_concurrency.lockutils [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.816126] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ca875bb-c351-4346-872a-6195eef3022a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.833124] env[62600]: DEBUG nova.virt.hardware [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 983.833379] env[62600]: DEBUG nova.virt.hardware [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 983.833543] env[62600]: DEBUG nova.virt.hardware [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 983.833733] env[62600]: DEBUG nova.virt.hardware [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 983.833883] env[62600]: DEBUG nova.virt.hardware [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 983.834056] env[62600]: DEBUG nova.virt.hardware [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 983.834316] env[62600]: DEBUG nova.virt.hardware [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 983.834487] env[62600]: DEBUG nova.virt.hardware [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 983.834660] env[62600]: DEBUG nova.virt.hardware [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 983.834826] env[62600]: DEBUG nova.virt.hardware [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 983.835012] env[62600]: DEBUG nova.virt.hardware [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 983.841274] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Reconfiguring VM to attach interface {{(pid=62600) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 983.844197] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c4fd0075-34e3-456c-8e19-1e7121458179 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.865738] env[62600]: DEBUG oslo_vmware.api [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 983.865738] env[62600]: value = "task-1222890" [ 983.865738] env[62600]: _type = "Task" [ 983.865738] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.877234] env[62600]: DEBUG oslo_vmware.api [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222890, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.908968] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05bdaa90-c47d-4d3f-9c53-8b8d26897a4e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.917021] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32655ccd-17ae-4a75-b9c7-9c7915f90bce {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.952484] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ddea7e-5996-4595-baad-a94a690dd78a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.959914] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2322fbdf-2386-4b68-b39b-196914f107ea {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.974555] env[62600]: DEBUG nova.compute.provider_tree [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.041977] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52a7b197-b799-ea74-1127-e5e299c0887b, 'name': SearchDatastore_Task, 'duration_secs': 0.009032} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.042782] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04e9ac93-1aa6-42ae-a55c-5af2cce416b5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.048131] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Waiting for the task: (returnval){ [ 984.048131] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52189959-2819-c7d8-a8ee-61afc3c096c4" [ 984.048131] env[62600]: _type = "Task" [ 984.048131] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.055860] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52189959-2819-c7d8-a8ee-61afc3c096c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.110494] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 984.110789] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 984.111069] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 984.111218] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 984.111397] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 984.111748] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 984.112196] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 984.112382] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 984.112615] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 984.112888] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 984.113276] env[62600]: DEBUG nova.virt.hardware [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 984.118940] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Reconfiguring VM instance instance-0000005a to detach disk 2000 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 984.119585] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-55b92364-fffb-40e8-b2cb-415cea29044d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.137859] env[62600]: DEBUG oslo_vmware.api [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 984.137859] env[62600]: value = "task-1222891" [ 984.137859] env[62600]: _type = "Task" [ 984.137859] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.146528] env[62600]: DEBUG oslo_vmware.api [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222891, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.264805] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.361676] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Acquiring lock "310a9093-4066-4bc3-af94-5d606860dbad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.361836] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Lock "310a9093-4066-4bc3-af94-5d606860dbad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.362095] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Acquiring lock "310a9093-4066-4bc3-af94-5d606860dbad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.362756] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Lock "310a9093-4066-4bc3-af94-5d606860dbad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.362961] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Lock "310a9093-4066-4bc3-af94-5d606860dbad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.365098] env[62600]: INFO nova.compute.manager [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Terminating instance [ 984.369685] env[62600]: DEBUG nova.compute.manager [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 984.369887] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 984.370642] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bfc577e-82ee-4343-9f87-93b93526900a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.378986] env[62600]: DEBUG oslo_vmware.api [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222890, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.381039] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 984.381272] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7933084a-fc8d-4b03-ad11-9a4443ab204c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.387991] env[62600]: DEBUG oslo_vmware.api [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Waiting for the task: (returnval){ [ 984.387991] env[62600]: value = "task-1222892" [ 984.387991] env[62600]: _type = "Task" [ 984.387991] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.395264] env[62600]: DEBUG oslo_vmware.api [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Task: {'id': task-1222892, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.478345] env[62600]: DEBUG nova.scheduler.client.report [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.559364] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52189959-2819-c7d8-a8ee-61afc3c096c4, 'name': SearchDatastore_Task, 'duration_secs': 0.011153} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.559636] env[62600]: DEBUG oslo_concurrency.lockutils [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.559902] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] e4bc0326-66d8-45e2-bcd8-11c4fab16e34/e4bc0326-66d8-45e2-bcd8-11c4fab16e34.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 984.560183] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bdbb1968-986d-4d8c-a090-f9d376454c16 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.566920] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Waiting for the task: (returnval){ [ 984.566920] env[62600]: value = "task-1222893" [ 984.566920] env[62600]: _type = "Task" [ 984.566920] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.574767] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': task-1222893, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.648629] env[62600]: DEBUG oslo_vmware.api [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222891, 'name': ReconfigVM_Task, 'duration_secs': 0.157091} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.648908] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Reconfigured VM instance instance-0000005a to detach disk 2000 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 984.649686] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a89f0e1b-be85-4896-8520-0e619a801ba0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.673610] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 02c6dac9-535b-436b-bbee-c8031cb5d7f5/02c6dac9-535b-436b-bbee-c8031cb5d7f5.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 984.673884] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9fd9cac8-65d9-4e72-88d9-527c7167188a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.691484] env[62600]: DEBUG oslo_vmware.api [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 984.691484] env[62600]: value = "task-1222894" [ 984.691484] env[62600]: _type = "Task" [ 984.691484] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.699731] env[62600]: DEBUG nova.compute.manager [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 984.701638] env[62600]: DEBUG oslo_vmware.api [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222894, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.727050] env[62600]: DEBUG nova.virt.hardware [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 984.727316] env[62600]: DEBUG nova.virt.hardware [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 984.727479] env[62600]: DEBUG nova.virt.hardware [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 984.727663] env[62600]: DEBUG nova.virt.hardware [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 984.727810] env[62600]: DEBUG nova.virt.hardware [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 984.728421] env[62600]: DEBUG nova.virt.hardware [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 984.728421] env[62600]: DEBUG nova.virt.hardware [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 984.728421] env[62600]: DEBUG nova.virt.hardware [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 984.728872] env[62600]: DEBUG nova.virt.hardware [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 984.728872] env[62600]: DEBUG nova.virt.hardware [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 984.728872] env[62600]: DEBUG nova.virt.hardware [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 984.729709] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c58d19c-79e1-4789-8067-ffb5f38d1bc4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.737643] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51cf05f1-3cc6-43f4-963b-f3dbdc9dbca0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.878315] env[62600]: DEBUG oslo_vmware.api [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222890, 'name': ReconfigVM_Task, 'duration_secs': 0.642428} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.878868] env[62600]: DEBUG oslo_concurrency.lockutils [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.879102] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Reconfigured VM to attach interface {{(pid=62600) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 984.898022] env[62600]: DEBUG oslo_vmware.api [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Task: {'id': task-1222892, 'name': PowerOffVM_Task, 'duration_secs': 0.216253} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.899021] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 984.899214] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 984.899483] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7798fac7-592b-41c8-b35f-57bed80067f2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.906328] env[62600]: DEBUG nova.compute.manager [req-822fc2e0-f5b1-4453-9092-c603c6a51e5c req-7ea91c68-7eb9-40d1-a71e-997184cb9bba service nova] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Received event network-vif-plugged-7af8f843-5635-4c50-bc15-f8356b10e6f7 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 984.906580] env[62600]: DEBUG oslo_concurrency.lockutils [req-822fc2e0-f5b1-4453-9092-c603c6a51e5c req-7ea91c68-7eb9-40d1-a71e-997184cb9bba service nova] Acquiring lock "aff91190-e19b-4f1c-8b47-9539360a4596-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.906784] env[62600]: DEBUG oslo_concurrency.lockutils [req-822fc2e0-f5b1-4453-9092-c603c6a51e5c req-7ea91c68-7eb9-40d1-a71e-997184cb9bba service nova] Lock "aff91190-e19b-4f1c-8b47-9539360a4596-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.907025] env[62600]: DEBUG oslo_concurrency.lockutils [req-822fc2e0-f5b1-4453-9092-c603c6a51e5c req-7ea91c68-7eb9-40d1-a71e-997184cb9bba service nova] Lock "aff91190-e19b-4f1c-8b47-9539360a4596-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.907393] env[62600]: DEBUG nova.compute.manager [req-822fc2e0-f5b1-4453-9092-c603c6a51e5c req-7ea91c68-7eb9-40d1-a71e-997184cb9bba service nova] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] No waiting events found dispatching network-vif-plugged-7af8f843-5635-4c50-bc15-f8356b10e6f7 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 984.907393] env[62600]: WARNING nova.compute.manager [req-822fc2e0-f5b1-4453-9092-c603c6a51e5c req-7ea91c68-7eb9-40d1-a71e-997184cb9bba service nova] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Received unexpected event network-vif-plugged-7af8f843-5635-4c50-bc15-f8356b10e6f7 for instance with vm_state building and task_state spawning. [ 984.984772] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.310s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.985359] env[62600]: DEBUG nova.compute.manager [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 984.988682] env[62600]: DEBUG nova.network.neutron [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Successfully updated port: 7af8f843-5635-4c50-bc15-f8356b10e6f7 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 984.989895] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.725s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.990079] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.990247] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62600) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 984.992180] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf667ce-6f40-489e-81bc-a01e53335278 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.995035] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 984.995434] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 984.995434] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Deleting the datastore file [datastore2] 310a9093-4066-4bc3-af94-5d606860dbad {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 984.996048] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ae2aa7a4-f6ae-43c3-87df-08e596c06963 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.005280] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8fbf86a-c5fb-400f-9502-6d3bd66d1a29 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.009626] env[62600]: DEBUG oslo_vmware.api [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Waiting for the task: (returnval){ [ 985.009626] env[62600]: value = "task-1222896" [ 985.009626] env[62600]: _type = "Task" [ 985.009626] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.023462] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-642fae37-d937-403d-9aaa-919b50802c32 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.029950] env[62600]: DEBUG oslo_vmware.api [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Task: {'id': task-1222896, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.035122] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a49f355-8483-4b09-b87a-cb62097cf055 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.069185] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179567MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=62600) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 985.069521] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.069664] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.079712] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': task-1222893, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.200970] env[62600]: DEBUG oslo_vmware.api [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222894, 'name': ReconfigVM_Task, 'duration_secs': 0.294617} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.201278] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 02c6dac9-535b-436b-bbee-c8031cb5d7f5/02c6dac9-535b-436b-bbee-c8031cb5d7f5.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 985.201570] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Updating instance '02c6dac9-535b-436b-bbee-c8031cb5d7f5' progress to 50 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 985.383879] env[62600]: DEBUG oslo_concurrency.lockutils [None req-04fefd50-b305-4356-bc01-61b435612cfa tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "interface-88cf521d-f4a3-493a-894d-c7f799dc0443-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.497489] env[62600]: DEBUG nova.compute.utils [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 985.499014] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "refresh_cache-aff91190-e19b-4f1c-8b47-9539360a4596" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.500164] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired lock "refresh_cache-aff91190-e19b-4f1c-8b47-9539360a4596" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.500164] env[62600]: DEBUG nova.network.neutron [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 985.501183] env[62600]: DEBUG nova.compute.manager [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 985.501383] env[62600]: DEBUG nova.network.neutron [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 985.520207] env[62600]: DEBUG oslo_vmware.api [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Task: {'id': task-1222896, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.268497} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.520474] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 985.520664] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 985.521124] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 985.521124] env[62600]: INFO nova.compute.manager [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Took 1.15 seconds to destroy the instance on the hypervisor. [ 985.521261] env[62600]: DEBUG oslo.service.loopingcall [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.521450] env[62600]: DEBUG nova.compute.manager [-] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 985.521545] env[62600]: DEBUG nova.network.neutron [-] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 985.556710] env[62600]: DEBUG nova.policy [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3e59998217a4b18a6f3f01142a5e440', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f91091f83ee4a2091507ca994e3d52f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 985.585390] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': task-1222893, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.561832} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.585809] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] e4bc0326-66d8-45e2-bcd8-11c4fab16e34/e4bc0326-66d8-45e2-bcd8-11c4fab16e34.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 985.586830] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 985.586830] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1b885d39-64ea-48cc-bd0a-6b6f81f19cfb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.595536] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Waiting for the task: (returnval){ [ 985.595536] env[62600]: value = "task-1222897" [ 985.595536] env[62600]: _type = "Task" [ 985.595536] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.606526] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': task-1222897, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.708193] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72c6b7f1-7400-43fa-a18c-9dd7f4683889 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.731194] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad87aead-8259-4e79-89b2-3b7e7ec7e299 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.751756] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Updating instance '02c6dac9-535b-436b-bbee-c8031cb5d7f5' progress to 67 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 985.823866] env[62600]: DEBUG nova.compute.manager [req-2b64a699-3123-4916-9bad-35b6cdfcc9b0 req-7122896e-eec2-42f2-a553-b3f97a0a4751 service nova] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Received event network-vif-deleted-038c3f9f-2b55-451c-abfe-33f1ca92935d {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.824053] env[62600]: INFO nova.compute.manager [req-2b64a699-3123-4916-9bad-35b6cdfcc9b0 req-7122896e-eec2-42f2-a553-b3f97a0a4751 service nova] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Neutron deleted interface 038c3f9f-2b55-451c-abfe-33f1ca92935d; detaching it from the instance and deleting it from the info cache [ 985.824285] env[62600]: DEBUG nova.network.neutron [req-2b64a699-3123-4916-9bad-35b6cdfcc9b0 req-7122896e-eec2-42f2-a553-b3f97a0a4751 service nova] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.849456] env[62600]: DEBUG nova.network.neutron [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Successfully created port: 4c6c4d03-b2f6-4b73-acb7-7640e7047873 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 986.001806] env[62600]: DEBUG nova.compute.manager [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 986.069239] env[62600]: DEBUG nova.network.neutron [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 986.086828] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Applying migration context for instance 02c6dac9-535b-436b-bbee-c8031cb5d7f5 as it has an incoming, in-progress migration 8d35c436-258f-4615-a3a2-eb29f7f59091. Migration status is migrating {{(pid=62600) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 986.088452] env[62600]: INFO nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Updating resource usage from migration 8d35c436-258f-4615-a3a2-eb29f7f59091 [ 986.106080] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': task-1222897, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.145202} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.106663] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 986.107456] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f2273a-54d5-4079-b112-6eac5bb80935 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.128639] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] e4bc0326-66d8-45e2-bcd8-11c4fab16e34/e4bc0326-66d8-45e2-bcd8-11c4fab16e34.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 986.132072] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance b967fb11-e70e-4e17-b769-38da581bd83b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.132243] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance d66c404d-8fd1-4fb7-a3b9-f21854c7e735 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.132375] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 1b856aff-66be-4fa5-b8b7-124ac24fb4d1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.132495] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance ab11cdd1-80a7-43c4-afac-24c7930238c1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.132950] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 248a5371-6ff7-4da7-ae97-d638bea0123a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.132950] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance c2f38d56-72fa-4bf0-a2cb-13d666571466 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.132950] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 88cf521d-f4a3-493a-894d-c7f799dc0443 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.132950] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 39deb498-6bf0-4f3b-932b-8068fc48271e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.133208] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 310a9093-4066-4bc3-af94-5d606860dbad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.133208] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Migration 8d35c436-258f-4615-a3a2-eb29f7f59091 is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 986.133277] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 02c6dac9-535b-436b-bbee-c8031cb5d7f5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.133367] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance e4bc0326-66d8-45e2-bcd8-11c4fab16e34 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.133479] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance aff91190-e19b-4f1c-8b47-9539360a4596 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.133587] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 5d7c2a17-9dee-407e-a7ee-6bb88f56905f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 986.133784] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 986.133919] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3264MB phys_disk=200GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 986.135996] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7e07a54-c49d-4e2f-9278-775e5e7617bb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.157259] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Waiting for the task: (returnval){ [ 986.157259] env[62600]: value = "task-1222898" [ 986.157259] env[62600]: _type = "Task" [ 986.157259] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.167380] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': task-1222898, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.274583] env[62600]: DEBUG nova.network.neutron [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Updating instance_info_cache with network_info: [{"id": "7af8f843-5635-4c50-bc15-f8356b10e6f7", "address": "fa:16:3e:fc:4b:8b", "network": {"id": "5ef67e8a-5613-4cdb-9f35-0a18720657ff", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1644754500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bab74e3ae78248909dbcd483abbb8da2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7af8f843-56", "ovs_interfaceid": "7af8f843-5635-4c50-bc15-f8356b10e6f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.296700] env[62600]: DEBUG nova.network.neutron [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Port 3753b6a6-f27d-4e1d-b801-91324a7e4e0e binding to destination host cpu-1 is already ACTIVE {{(pid=62600) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 986.298959] env[62600]: DEBUG nova.network.neutron [-] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.326945] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f02e278e-d6a3-45e4-a555-8b2db80a1d99 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.338216] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c63b64b-0b66-48e5-82db-814932f5ef71 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.363824] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a27f06ba-dabe-46fe-a760-ac02bbc91dcc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.377424] env[62600]: DEBUG nova.compute.manager [req-2b64a699-3123-4916-9bad-35b6cdfcc9b0 req-7122896e-eec2-42f2-a553-b3f97a0a4751 service nova] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Detach interface failed, port_id=038c3f9f-2b55-451c-abfe-33f1ca92935d, reason: Instance 310a9093-4066-4bc3-af94-5d606860dbad could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 986.382595] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c22c3b-e826-41d2-8331-463550ea129a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.413882] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b5aedf-2479-418c-ae8f-1085bf65e5b1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.424342] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-425baf75-26ee-4233-ab56-c690f95900cd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.438489] env[62600]: DEBUG nova.compute.provider_tree [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.667915] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': task-1222898, 'name': ReconfigVM_Task, 'duration_secs': 0.250009} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.668212] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Reconfigured VM instance instance-0000005c to attach disk [datastore2] e4bc0326-66d8-45e2-bcd8-11c4fab16e34/e4bc0326-66d8-45e2-bcd8-11c4fab16e34.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 986.668902] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-406bf094-9e19-44a1-91d0-f65918652961 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.676373] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Waiting for the task: (returnval){ [ 986.676373] env[62600]: value = "task-1222899" [ 986.676373] env[62600]: _type = "Task" [ 986.676373] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.683879] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': task-1222899, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.777716] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Releasing lock "refresh_cache-aff91190-e19b-4f1c-8b47-9539360a4596" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.778098] env[62600]: DEBUG nova.compute.manager [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Instance network_info: |[{"id": "7af8f843-5635-4c50-bc15-f8356b10e6f7", "address": "fa:16:3e:fc:4b:8b", "network": {"id": "5ef67e8a-5613-4cdb-9f35-0a18720657ff", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1644754500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bab74e3ae78248909dbcd483abbb8da2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7af8f843-56", "ovs_interfaceid": "7af8f843-5635-4c50-bc15-f8356b10e6f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 986.778597] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:4b:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3734b156-0f7d-4721-b23c-d000412ec2eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7af8f843-5635-4c50-bc15-f8356b10e6f7', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 986.786759] env[62600]: DEBUG oslo.service.loopingcall [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.786994] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 986.787271] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e07d6e9c-3351-48a5-bd47-b57fa0a41645 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.806254] env[62600]: INFO nova.compute.manager [-] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Took 1.28 seconds to deallocate network for instance. [ 986.813873] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 986.813873] env[62600]: value = "task-1222900" [ 986.813873] env[62600]: _type = "Task" [ 986.813873] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.822097] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222900, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.919502] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "interface-88cf521d-f4a3-493a-894d-c7f799dc0443-0bc0b585-65b4-4ebf-81be-bbc628771202" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.919841] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "interface-88cf521d-f4a3-493a-894d-c7f799dc0443-0bc0b585-65b4-4ebf-81be-bbc628771202" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.920316] env[62600]: DEBUG nova.objects.instance [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lazy-loading 'flavor' on Instance uuid 88cf521d-f4a3-493a-894d-c7f799dc0443 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 986.935386] env[62600]: DEBUG nova.compute.manager [req-7cdd7cb1-4e68-4137-8c25-f6393600aff2 req-bfbf6ee1-5735-4388-8e4c-a389821020ff service nova] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Received event network-changed-7af8f843-5635-4c50-bc15-f8356b10e6f7 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 986.935608] env[62600]: DEBUG nova.compute.manager [req-7cdd7cb1-4e68-4137-8c25-f6393600aff2 req-bfbf6ee1-5735-4388-8e4c-a389821020ff service nova] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Refreshing instance network info cache due to event network-changed-7af8f843-5635-4c50-bc15-f8356b10e6f7. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 986.935772] env[62600]: DEBUG oslo_concurrency.lockutils [req-7cdd7cb1-4e68-4137-8c25-f6393600aff2 req-bfbf6ee1-5735-4388-8e4c-a389821020ff service nova] Acquiring lock "refresh_cache-aff91190-e19b-4f1c-8b47-9539360a4596" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.935969] env[62600]: DEBUG oslo_concurrency.lockutils [req-7cdd7cb1-4e68-4137-8c25-f6393600aff2 req-bfbf6ee1-5735-4388-8e4c-a389821020ff service nova] Acquired lock "refresh_cache-aff91190-e19b-4f1c-8b47-9539360a4596" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.936087] env[62600]: DEBUG nova.network.neutron [req-7cdd7cb1-4e68-4137-8c25-f6393600aff2 req-bfbf6ee1-5735-4388-8e4c-a389821020ff service nova] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Refreshing network info cache for port 7af8f843-5635-4c50-bc15-f8356b10e6f7 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 986.941118] env[62600]: DEBUG nova.scheduler.client.report [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 987.012828] env[62600]: DEBUG nova.compute.manager [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 987.039847] env[62600]: DEBUG nova.virt.hardware [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 987.040137] env[62600]: DEBUG nova.virt.hardware [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 987.040297] env[62600]: DEBUG nova.virt.hardware [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 987.040491] env[62600]: DEBUG nova.virt.hardware [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 987.040650] env[62600]: DEBUG nova.virt.hardware [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 987.040808] env[62600]: DEBUG nova.virt.hardware [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 987.041137] env[62600]: DEBUG nova.virt.hardware [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 987.041316] env[62600]: DEBUG nova.virt.hardware [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 987.041494] env[62600]: DEBUG nova.virt.hardware [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 987.041665] env[62600]: DEBUG nova.virt.hardware [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 987.041845] env[62600]: DEBUG nova.virt.hardware [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.042744] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee8b4467-fa60-4c94-93af-223a5b435f5e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.051806] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a4f472-ac11-4b0f-a7cf-23deeafa2f45 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.186351] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': task-1222899, 'name': Rename_Task, 'duration_secs': 0.125656} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.186651] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 987.186895] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aa26798e-7d95-452f-8385-c98aa86ff06d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.193313] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Waiting for the task: (returnval){ [ 987.193313] env[62600]: value = "task-1222901" [ 987.193313] env[62600]: _type = "Task" [ 987.193313] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.201111] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': task-1222901, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.316229] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.326198] env[62600]: DEBUG oslo_concurrency.lockutils [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "02c6dac9-535b-436b-bbee-c8031cb5d7f5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.326436] env[62600]: DEBUG oslo_concurrency.lockutils [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "02c6dac9-535b-436b-bbee-c8031cb5d7f5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.328029] env[62600]: DEBUG oslo_concurrency.lockutils [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "02c6dac9-535b-436b-bbee-c8031cb5d7f5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.334256] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222900, 'name': CreateVM_Task, 'duration_secs': 0.355008} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.335844] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 987.335844] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.335844] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.336759] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 987.336759] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46f4e3b3-9eed-4c2c-b55a-108a2946aa70 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.342226] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 987.342226] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c4d6bf-4a0e-6327-5402-b1e9f519637a" [ 987.342226] env[62600]: _type = "Task" [ 987.342226] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.353151] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c4d6bf-4a0e-6327-5402-b1e9f519637a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.377719] env[62600]: DEBUG nova.network.neutron [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Successfully updated port: 4c6c4d03-b2f6-4b73-acb7-7640e7047873 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 987.446298] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62600) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 987.449020] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.377s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.449020] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.131s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.449020] env[62600]: DEBUG nova.objects.instance [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Lazy-loading 'resources' on Instance uuid 310a9093-4066-4bc3-af94-5d606860dbad {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.530927] env[62600]: DEBUG nova.objects.instance [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lazy-loading 'pci_requests' on Instance uuid 88cf521d-f4a3-493a-894d-c7f799dc0443 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.706947] env[62600]: DEBUG oslo_vmware.api [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': task-1222901, 'name': PowerOnVM_Task, 'duration_secs': 0.439576} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.707460] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 987.707800] env[62600]: INFO nova.compute.manager [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Took 5.31 seconds to spawn the instance on the hypervisor. [ 987.708131] env[62600]: DEBUG nova.compute.manager [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 987.710204] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae5c1e47-f894-4501-be88-52d2d7acedac {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.853692] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c4d6bf-4a0e-6327-5402-b1e9f519637a, 'name': SearchDatastore_Task, 'duration_secs': 0.021672} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.853989] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.854269] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 987.854528] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.854705] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.854887] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 987.855195] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a3d4c74f-e7d3-4e32-9f2b-fba38265aab4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.863465] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 987.863646] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 987.864379] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abdb313a-2cff-46a3-a234-6930bb03448d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.869389] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 987.869389] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c72467-03c6-91b8-5b85-9988971144e4" [ 987.869389] env[62600]: _type = "Task" [ 987.869389] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.877307] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c72467-03c6-91b8-5b85-9988971144e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.880706] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "refresh_cache-5d7c2a17-9dee-407e-a7ee-6bb88f56905f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.880953] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "refresh_cache-5d7c2a17-9dee-407e-a7ee-6bb88f56905f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.881046] env[62600]: DEBUG nova.network.neutron [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 987.893345] env[62600]: DEBUG nova.network.neutron [req-7cdd7cb1-4e68-4137-8c25-f6393600aff2 req-bfbf6ee1-5735-4388-8e4c-a389821020ff service nova] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Updated VIF entry in instance network info cache for port 7af8f843-5635-4c50-bc15-f8356b10e6f7. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 987.893689] env[62600]: DEBUG nova.network.neutron [req-7cdd7cb1-4e68-4137-8c25-f6393600aff2 req-bfbf6ee1-5735-4388-8e4c-a389821020ff service nova] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Updating instance_info_cache with network_info: [{"id": "7af8f843-5635-4c50-bc15-f8356b10e6f7", "address": "fa:16:3e:fc:4b:8b", "network": {"id": "5ef67e8a-5613-4cdb-9f35-0a18720657ff", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1644754500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bab74e3ae78248909dbcd483abbb8da2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7af8f843-56", "ovs_interfaceid": "7af8f843-5635-4c50-bc15-f8356b10e6f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.033461] env[62600]: DEBUG nova.objects.base [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Object Instance<88cf521d-f4a3-493a-894d-c7f799dc0443> lazy-loaded attributes: flavor,pci_requests {{(pid=62600) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 988.033739] env[62600]: DEBUG nova.network.neutron [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 988.107316] env[62600]: DEBUG nova.policy [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fbd1b2fb34d841359ada8fc44bec2986', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b028450f2da445fb83e37adfc86bba68', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 988.114842] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-854fb902-368d-4986-a392-50eb28c1963b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.123070] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df208b2b-b7b8-402f-b771-ecaf5fd502cf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.155736] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7267a4-cdc6-48af-967e-a44136150418 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.166863] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bea0a652-070c-4bbc-99bb-7c9bd20bb0b2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.181670] env[62600]: DEBUG nova.compute.provider_tree [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 988.227383] env[62600]: INFO nova.compute.manager [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Took 13.98 seconds to build instance. [ 988.365289] env[62600]: DEBUG oslo_concurrency.lockutils [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "refresh_cache-02c6dac9-535b-436b-bbee-c8031cb5d7f5" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.365543] env[62600]: DEBUG oslo_concurrency.lockutils [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "refresh_cache-02c6dac9-535b-436b-bbee-c8031cb5d7f5" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.365731] env[62600]: DEBUG nova.network.neutron [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 988.380705] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c72467-03c6-91b8-5b85-9988971144e4, 'name': SearchDatastore_Task, 'duration_secs': 0.008704} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.381687] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cbbe48df-869d-4faf-b783-fb524cb818ef {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.387734] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 988.387734] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52882bde-0bbc-581c-ea3b-f33f516bb6c6" [ 988.387734] env[62600]: _type = "Task" [ 988.387734] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.395984] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52882bde-0bbc-581c-ea3b-f33f516bb6c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.396448] env[62600]: DEBUG oslo_concurrency.lockutils [req-7cdd7cb1-4e68-4137-8c25-f6393600aff2 req-bfbf6ee1-5735-4388-8e4c-a389821020ff service nova] Releasing lock "refresh_cache-aff91190-e19b-4f1c-8b47-9539360a4596" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.412582] env[62600]: DEBUG nova.network.neutron [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 988.541968] env[62600]: DEBUG nova.network.neutron [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Updating instance_info_cache with network_info: [{"id": "4c6c4d03-b2f6-4b73-acb7-7640e7047873", "address": "fa:16:3e:09:6e:66", "network": {"id": "1d2052cf-5be1-48e7-887a-18f367b7823f", "bridge": "br-int", "label": "tempest-ServersTestJSON-106099679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f91091f83ee4a2091507ca994e3d52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c6c4d03-b2", "ovs_interfaceid": "4c6c4d03-b2f6-4b73-acb7-7640e7047873", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.684461] env[62600]: DEBUG nova.scheduler.client.report [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 988.730189] env[62600]: DEBUG oslo_concurrency.lockutils [None req-05c01f3b-30ac-4ef8-8d74-fb90fb1980d9 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Lock "e4bc0326-66d8-45e2-bcd8-11c4fab16e34" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.487s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.846676] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.846896] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.898960] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52882bde-0bbc-581c-ea3b-f33f516bb6c6, 'name': SearchDatastore_Task, 'duration_secs': 0.0103} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.899313] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.899610] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] aff91190-e19b-4f1c-8b47-9539360a4596/aff91190-e19b-4f1c-8b47-9539360a4596.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 988.899885] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c997b389-82f5-4829-9a80-a13077748035 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.909379] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 988.909379] env[62600]: value = "task-1222902" [ 988.909379] env[62600]: _type = "Task" [ 988.909379] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.918398] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222902, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.967205] env[62600]: DEBUG nova.compute.manager [None req-e8bb7539-86e1-43ae-97b2-f3ba36ed539b tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.968263] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff029a6-66aa-4c1e-9db7-15aad99bcc33 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.010349] env[62600]: DEBUG nova.compute.manager [req-a69bdb64-e9fb-4523-8a15-c57128d4c404 req-a4462d5f-0c77-42d2-b031-dc2c7787727e service nova] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Received event network-vif-plugged-4c6c4d03-b2f6-4b73-acb7-7640e7047873 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.010567] env[62600]: DEBUG oslo_concurrency.lockutils [req-a69bdb64-e9fb-4523-8a15-c57128d4c404 req-a4462d5f-0c77-42d2-b031-dc2c7787727e service nova] Acquiring lock "5d7c2a17-9dee-407e-a7ee-6bb88f56905f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.010781] env[62600]: DEBUG oslo_concurrency.lockutils [req-a69bdb64-e9fb-4523-8a15-c57128d4c404 req-a4462d5f-0c77-42d2-b031-dc2c7787727e service nova] Lock "5d7c2a17-9dee-407e-a7ee-6bb88f56905f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.010987] env[62600]: DEBUG oslo_concurrency.lockutils [req-a69bdb64-e9fb-4523-8a15-c57128d4c404 req-a4462d5f-0c77-42d2-b031-dc2c7787727e service nova] Lock "5d7c2a17-9dee-407e-a7ee-6bb88f56905f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.011143] env[62600]: DEBUG nova.compute.manager [req-a69bdb64-e9fb-4523-8a15-c57128d4c404 req-a4462d5f-0c77-42d2-b031-dc2c7787727e service nova] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] No waiting events found dispatching network-vif-plugged-4c6c4d03-b2f6-4b73-acb7-7640e7047873 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 989.011314] env[62600]: WARNING nova.compute.manager [req-a69bdb64-e9fb-4523-8a15-c57128d4c404 req-a4462d5f-0c77-42d2-b031-dc2c7787727e service nova] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Received unexpected event network-vif-plugged-4c6c4d03-b2f6-4b73-acb7-7640e7047873 for instance with vm_state building and task_state spawning. [ 989.011478] env[62600]: DEBUG nova.compute.manager [req-a69bdb64-e9fb-4523-8a15-c57128d4c404 req-a4462d5f-0c77-42d2-b031-dc2c7787727e service nova] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Received event network-changed-4c6c4d03-b2f6-4b73-acb7-7640e7047873 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.011646] env[62600]: DEBUG nova.compute.manager [req-a69bdb64-e9fb-4523-8a15-c57128d4c404 req-a4462d5f-0c77-42d2-b031-dc2c7787727e service nova] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Refreshing instance network info cache due to event network-changed-4c6c4d03-b2f6-4b73-acb7-7640e7047873. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 989.011872] env[62600]: DEBUG oslo_concurrency.lockutils [req-a69bdb64-e9fb-4523-8a15-c57128d4c404 req-a4462d5f-0c77-42d2-b031-dc2c7787727e service nova] Acquiring lock "refresh_cache-5d7c2a17-9dee-407e-a7ee-6bb88f56905f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.046095] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "refresh_cache-5d7c2a17-9dee-407e-a7ee-6bb88f56905f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.046495] env[62600]: DEBUG nova.compute.manager [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Instance network_info: |[{"id": "4c6c4d03-b2f6-4b73-acb7-7640e7047873", "address": "fa:16:3e:09:6e:66", "network": {"id": "1d2052cf-5be1-48e7-887a-18f367b7823f", "bridge": "br-int", "label": "tempest-ServersTestJSON-106099679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f91091f83ee4a2091507ca994e3d52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c6c4d03-b2", "ovs_interfaceid": "4c6c4d03-b2f6-4b73-acb7-7640e7047873", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 989.047300] env[62600]: DEBUG oslo_concurrency.lockutils [req-a69bdb64-e9fb-4523-8a15-c57128d4c404 req-a4462d5f-0c77-42d2-b031-dc2c7787727e service nova] Acquired lock "refresh_cache-5d7c2a17-9dee-407e-a7ee-6bb88f56905f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.047568] env[62600]: DEBUG nova.network.neutron [req-a69bdb64-e9fb-4523-8a15-c57128d4c404 req-a4462d5f-0c77-42d2-b031-dc2c7787727e service nova] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Refreshing network info cache for port 4c6c4d03-b2f6-4b73-acb7-7640e7047873 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 989.049509] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:6e:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '028bae2d-fe6c-4207-b4a3-3fab45fbf1d6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4c6c4d03-b2f6-4b73-acb7-7640e7047873', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 989.057440] env[62600]: DEBUG oslo.service.loopingcall [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.058469] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 989.058724] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c1412a93-7c2f-4d1a-9f29-231ab45c6d38 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.083335] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 989.083335] env[62600]: value = "task-1222903" [ 989.083335] env[62600]: _type = "Task" [ 989.083335] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.093231] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222903, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.115170] env[62600]: DEBUG nova.network.neutron [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Updating instance_info_cache with network_info: [{"id": "3753b6a6-f27d-4e1d-b801-91324a7e4e0e", "address": "fa:16:3e:fd:6e:a3", "network": {"id": "95818650-8227-4d11-9d14-007f06c07790", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1738152115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73e664d178f7484a9f4741b4d9450e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3753b6a6-f2", "ovs_interfaceid": "3753b6a6-f27d-4e1d-b801-91324a7e4e0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.116720] env[62600]: DEBUG oslo_concurrency.lockutils [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Acquiring lock "e4bc0326-66d8-45e2-bcd8-11c4fab16e34" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.116981] env[62600]: DEBUG oslo_concurrency.lockutils [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Lock "e4bc0326-66d8-45e2-bcd8-11c4fab16e34" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.117266] env[62600]: DEBUG oslo_concurrency.lockutils [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Acquiring lock "e4bc0326-66d8-45e2-bcd8-11c4fab16e34-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.117518] env[62600]: DEBUG oslo_concurrency.lockutils [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Lock "e4bc0326-66d8-45e2-bcd8-11c4fab16e34-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.117698] env[62600]: DEBUG oslo_concurrency.lockutils [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Lock "e4bc0326-66d8-45e2-bcd8-11c4fab16e34-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.119958] env[62600]: INFO nova.compute.manager [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Terminating instance [ 989.121729] env[62600]: DEBUG oslo_concurrency.lockutils [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Acquiring lock "refresh_cache-e4bc0326-66d8-45e2-bcd8-11c4fab16e34" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.121897] env[62600]: DEBUG oslo_concurrency.lockutils [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Acquired lock "refresh_cache-e4bc0326-66d8-45e2-bcd8-11c4fab16e34" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.122183] env[62600]: DEBUG nova.network.neutron [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 989.189673] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.742s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.221283] env[62600]: INFO nova.scheduler.client.report [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Deleted allocations for instance 310a9093-4066-4bc3-af94-5d606860dbad [ 989.359021] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 989.359021] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Starting heal instance info cache {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 989.420545] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222902, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.446753} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.420847] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] aff91190-e19b-4f1c-8b47-9539360a4596/aff91190-e19b-4f1c-8b47-9539360a4596.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 989.421103] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 989.421412] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-90331279-7cbc-4527-894b-2e016ed17efd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.429778] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 989.429778] env[62600]: value = "task-1222904" [ 989.429778] env[62600]: _type = "Task" [ 989.429778] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.439295] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222904, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.479271] env[62600]: INFO nova.compute.manager [None req-e8bb7539-86e1-43ae-97b2-f3ba36ed539b tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] instance snapshotting [ 989.480181] env[62600]: DEBUG nova.objects.instance [None req-e8bb7539-86e1-43ae-97b2-f3ba36ed539b tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Lazy-loading 'flavor' on Instance uuid e4bc0326-66d8-45e2-bcd8-11c4fab16e34 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 989.521959] env[62600]: DEBUG nova.compute.manager [req-515ede74-cf40-40eb-8ab7-9c8d3e0168fd req-a09617eb-45fd-407e-a7db-8a957f56cce3 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Received event network-vif-plugged-0bc0b585-65b4-4ebf-81be-bbc628771202 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.522287] env[62600]: DEBUG oslo_concurrency.lockutils [req-515ede74-cf40-40eb-8ab7-9c8d3e0168fd req-a09617eb-45fd-407e-a7db-8a957f56cce3 service nova] Acquiring lock "88cf521d-f4a3-493a-894d-c7f799dc0443-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.522518] env[62600]: DEBUG oslo_concurrency.lockutils [req-515ede74-cf40-40eb-8ab7-9c8d3e0168fd req-a09617eb-45fd-407e-a7db-8a957f56cce3 service nova] Lock "88cf521d-f4a3-493a-894d-c7f799dc0443-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.522691] env[62600]: DEBUG oslo_concurrency.lockutils [req-515ede74-cf40-40eb-8ab7-9c8d3e0168fd req-a09617eb-45fd-407e-a7db-8a957f56cce3 service nova] Lock "88cf521d-f4a3-493a-894d-c7f799dc0443-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.522870] env[62600]: DEBUG nova.compute.manager [req-515ede74-cf40-40eb-8ab7-9c8d3e0168fd req-a09617eb-45fd-407e-a7db-8a957f56cce3 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] No waiting events found dispatching network-vif-plugged-0bc0b585-65b4-4ebf-81be-bbc628771202 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 989.523052] env[62600]: WARNING nova.compute.manager [req-515ede74-cf40-40eb-8ab7-9c8d3e0168fd req-a09617eb-45fd-407e-a7db-8a957f56cce3 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Received unexpected event network-vif-plugged-0bc0b585-65b4-4ebf-81be-bbc628771202 for instance with vm_state active and task_state None. [ 989.597267] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222903, 'name': CreateVM_Task, 'duration_secs': 0.378416} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.597454] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 989.598117] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.598292] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.598610] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 989.599117] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de96a813-c54e-4a91-872d-bcadd1e9e6af {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.603942] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 989.603942] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f20b85-a28a-8d2a-aab0-54e18a745171" [ 989.603942] env[62600]: _type = "Task" [ 989.603942] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.608122] env[62600]: DEBUG nova.network.neutron [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Successfully updated port: 0bc0b585-65b4-4ebf-81be-bbc628771202 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 989.614847] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f20b85-a28a-8d2a-aab0-54e18a745171, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.620872] env[62600]: DEBUG oslo_concurrency.lockutils [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "refresh_cache-02c6dac9-535b-436b-bbee-c8031cb5d7f5" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.646980] env[62600]: DEBUG nova.network.neutron [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 989.711683] env[62600]: DEBUG nova.network.neutron [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.730510] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ec8bb24c-d942-4621-954f-1cd0d4086fd8 tempest-ServerTagsTestJSON-710910405 tempest-ServerTagsTestJSON-710910405-project-member] Lock "310a9093-4066-4bc3-af94-5d606860dbad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.368s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.769652] env[62600]: DEBUG nova.network.neutron [req-a69bdb64-e9fb-4523-8a15-c57128d4c404 req-a4462d5f-0c77-42d2-b031-dc2c7787727e service nova] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Updated VIF entry in instance network info cache for port 4c6c4d03-b2f6-4b73-acb7-7640e7047873. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 989.770090] env[62600]: DEBUG nova.network.neutron [req-a69bdb64-e9fb-4523-8a15-c57128d4c404 req-a4462d5f-0c77-42d2-b031-dc2c7787727e service nova] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Updating instance_info_cache with network_info: [{"id": "4c6c4d03-b2f6-4b73-acb7-7640e7047873", "address": "fa:16:3e:09:6e:66", "network": {"id": "1d2052cf-5be1-48e7-887a-18f367b7823f", "bridge": "br-int", "label": "tempest-ServersTestJSON-106099679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f91091f83ee4a2091507ca994e3d52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c6c4d03-b2", "ovs_interfaceid": "4c6c4d03-b2f6-4b73-acb7-7640e7047873", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.908629] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "refresh_cache-d66c404d-8fd1-4fb7-a3b9-f21854c7e735" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.908859] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquired lock "refresh_cache-d66c404d-8fd1-4fb7-a3b9-f21854c7e735" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.908948] env[62600]: DEBUG nova.network.neutron [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Forcefully refreshing network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 989.939829] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222904, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065066} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.940115] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 989.940906] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c5b6f2-d7eb-4631-a660-bbc0bcaacd28 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.963407] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] aff91190-e19b-4f1c-8b47-9539360a4596/aff91190-e19b-4f1c-8b47-9539360a4596.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 989.963746] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-823568ea-fe3b-4ede-9634-994689906b54 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.986035] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 989.986035] env[62600]: value = "task-1222905" [ 989.986035] env[62600]: _type = "Task" [ 989.986035] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.987080] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c474a634-d660-44ff-b5e0-57e115592ca3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.998365] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222905, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.009470] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-073f3ebe-a88a-40d9-ba10-088c4cf3d1a8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.111345] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.111651] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.111860] env[62600]: DEBUG nova.network.neutron [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 990.117584] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f20b85-a28a-8d2a-aab0-54e18a745171, 'name': SearchDatastore_Task, 'duration_secs': 0.030746} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.118126] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.118382] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 990.118624] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.118773] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.118957] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 990.119236] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-15301ea4-16bd-4e4d-8b66-dc3111412d6e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.132070] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 990.132501] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 990.135450] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba02957a-4632-4ce5-b5b8-8c0de62efaea {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.145034] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 990.145034] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5285ea51-fb59-c8f8-b330-943f137d3cac" [ 990.145034] env[62600]: _type = "Task" [ 990.145034] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.153995] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5285ea51-fb59-c8f8-b330-943f137d3cac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.155506] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c09c0a41-9f11-4765-947e-24dec61a4d3a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.174475] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e9b2e46-60b1-4d43-8ced-d51d7a7a8a6a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.182393] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Updating instance '02c6dac9-535b-436b-bbee-c8031cb5d7f5' progress to 83 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 990.214725] env[62600]: DEBUG oslo_concurrency.lockutils [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Releasing lock "refresh_cache-e4bc0326-66d8-45e2-bcd8-11c4fab16e34" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.215178] env[62600]: DEBUG nova.compute.manager [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 990.215402] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 990.216348] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d6da97-522e-42a9-8f8a-4acb8ac5d9b5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.225377] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 990.226488] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2378cd39-4c58-492e-af7f-ffc0ffdc09e2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.235316] env[62600]: DEBUG oslo_vmware.api [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Waiting for the task: (returnval){ [ 990.235316] env[62600]: value = "task-1222906" [ 990.235316] env[62600]: _type = "Task" [ 990.235316] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.245599] env[62600]: DEBUG oslo_vmware.api [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': task-1222906, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.272826] env[62600]: DEBUG oslo_concurrency.lockutils [req-a69bdb64-e9fb-4523-8a15-c57128d4c404 req-a4462d5f-0c77-42d2-b031-dc2c7787727e service nova] Releasing lock "refresh_cache-5d7c2a17-9dee-407e-a7ee-6bb88f56905f" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.497450] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222905, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.522304] env[62600]: DEBUG nova.compute.manager [None req-e8bb7539-86e1-43ae-97b2-f3ba36ed539b tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Instance disappeared during snapshot {{(pid=62600) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 990.654204] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5285ea51-fb59-c8f8-b330-943f137d3cac, 'name': SearchDatastore_Task, 'duration_secs': 0.010946} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.654952] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aee62658-921b-4fb2-94b4-1d09596ac974 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.663726] env[62600]: WARNING nova.network.neutron [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] 8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb already exists in list: networks containing: ['8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb']. ignoring it [ 990.663941] env[62600]: WARNING nova.network.neutron [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] 8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb already exists in list: networks containing: ['8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb']. ignoring it [ 990.673304] env[62600]: DEBUG nova.compute.manager [None req-e8bb7539-86e1-43ae-97b2-f3ba36ed539b tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Found 0 images (rotation: 2) {{(pid=62600) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 990.678019] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 990.678019] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b4df35-0399-5aca-37e8-8e422aa7c9bc" [ 990.678019] env[62600]: _type = "Task" [ 990.678019] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.683664] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b4df35-0399-5aca-37e8-8e422aa7c9bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.689051] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 990.689051] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-62db5668-31b9-4517-85ca-206642aaacbf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.696775] env[62600]: DEBUG oslo_vmware.api [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 990.696775] env[62600]: value = "task-1222907" [ 990.696775] env[62600]: _type = "Task" [ 990.696775] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.705240] env[62600]: DEBUG oslo_vmware.api [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222907, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.747029] env[62600]: DEBUG oslo_vmware.api [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': task-1222906, 'name': PowerOffVM_Task, 'duration_secs': 0.220299} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.747029] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 990.747278] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 990.747457] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e23189db-9e0b-4894-8f85-ce5fb82373c1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.776736] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 990.777011] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 990.777206] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Deleting the datastore file [datastore2] e4bc0326-66d8-45e2-bcd8-11c4fab16e34 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 990.777486] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7cbebc29-46f1-4f1b-ac1a-ccd4315e19ee {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.789030] env[62600]: DEBUG oslo_vmware.api [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Waiting for the task: (returnval){ [ 990.789030] env[62600]: value = "task-1222909" [ 990.789030] env[62600]: _type = "Task" [ 990.789030] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.793281] env[62600]: DEBUG oslo_vmware.api [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': task-1222909, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.004179] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222905, 'name': ReconfigVM_Task, 'duration_secs': 0.586076} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.004516] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Reconfigured VM instance instance-0000005d to attach disk [datastore1] aff91190-e19b-4f1c-8b47-9539360a4596/aff91190-e19b-4f1c-8b47-9539360a4596.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 991.005155] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d0d27454-0320-4d76-9392-3ad6780774af {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.016846] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 991.016846] env[62600]: value = "task-1222910" [ 991.016846] env[62600]: _type = "Task" [ 991.016846] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.034241] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222910, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.148018] env[62600]: DEBUG nova.network.neutron [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Updating instance_info_cache with network_info: [{"id": "57606b61-13bc-4562-b28d-d61f215809e0", "address": "fa:16:3e:53:05:09", "network": {"id": "036dc114-6ea0-46cb-a958-4a6137e7e85f", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-728627542-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48e1772c9a6c4671ab89df2e8b40cae1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57606b61-13", "ovs_interfaceid": "57606b61-13bc-4562-b28d-d61f215809e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.187315] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b4df35-0399-5aca-37e8-8e422aa7c9bc, 'name': SearchDatastore_Task, 'duration_secs': 0.009514} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.187624] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.187901] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 5d7c2a17-9dee-407e-a7ee-6bb88f56905f/5d7c2a17-9dee-407e-a7ee-6bb88f56905f.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 991.188204] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b808219a-e60b-4d53-abfc-3bcf34e640f9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.197681] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 991.197681] env[62600]: value = "task-1222911" [ 991.197681] env[62600]: _type = "Task" [ 991.197681] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.209283] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222911, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.212276] env[62600]: DEBUG oslo_vmware.api [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222907, 'name': PowerOnVM_Task, 'duration_secs': 0.389132} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.212528] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 991.212756] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-419afe10-c645-4d1c-bd23-7cd1d27d890a tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Updating instance '02c6dac9-535b-436b-bbee-c8031cb5d7f5' progress to 100 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 991.301071] env[62600]: DEBUG oslo_vmware.api [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Task: {'id': task-1222909, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.099111} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.302815] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 991.302815] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 991.302815] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 991.305009] env[62600]: INFO nova.compute.manager [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Took 1.09 seconds to destroy the instance on the hypervisor. [ 991.305009] env[62600]: DEBUG oslo.service.loopingcall [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.305349] env[62600]: DEBUG nova.compute.manager [-] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 991.308479] env[62600]: DEBUG nova.network.neutron [-] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 991.330515] env[62600]: DEBUG nova.network.neutron [-] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 991.337914] env[62600]: DEBUG nova.network.neutron [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Updating instance_info_cache with network_info: [{"id": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "address": "fa:16:3e:67:24:01", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8aca44-97", "ovs_interfaceid": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "24f6c3e5-a847-44bd-ba62-c4d66a2489d8", "address": "fa:16:3e:e1:5b:18", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24f6c3e5-a8", "ovs_interfaceid": "24f6c3e5-a847-44bd-ba62-c4d66a2489d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0bc0b585-65b4-4ebf-81be-bbc628771202", "address": "fa:16:3e:da:a7:96", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bc0b585-65", "ovs_interfaceid": "0bc0b585-65b4-4ebf-81be-bbc628771202", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.529364] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222910, 'name': Rename_Task, 'duration_secs': 0.170618} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.529561] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 991.529828] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-43aebdec-b87c-4f7e-9708-cb9ccc681f17 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.540491] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 991.540491] env[62600]: value = "task-1222912" [ 991.540491] env[62600]: _type = "Task" [ 991.540491] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.549400] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222912, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.556481] env[62600]: DEBUG nova.compute.manager [req-9ec72425-a779-4345-bff1-51922ce6feef req-cd64fb93-e846-4f0b-91dd-845b27e18234 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Received event network-changed-0bc0b585-65b4-4ebf-81be-bbc628771202 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.556481] env[62600]: DEBUG nova.compute.manager [req-9ec72425-a779-4345-bff1-51922ce6feef req-cd64fb93-e846-4f0b-91dd-845b27e18234 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Refreshing instance network info cache due to event network-changed-0bc0b585-65b4-4ebf-81be-bbc628771202. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 991.556481] env[62600]: DEBUG oslo_concurrency.lockutils [req-9ec72425-a779-4345-bff1-51922ce6feef req-cd64fb93-e846-4f0b-91dd-845b27e18234 service nova] Acquiring lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.649605] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Releasing lock "refresh_cache-d66c404d-8fd1-4fb7-a3b9-f21854c7e735" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.649605] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Updated the network info_cache for instance {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 991.712814] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222911, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.514133} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.713183] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 5d7c2a17-9dee-407e-a7ee-6bb88f56905f/5d7c2a17-9dee-407e-a7ee-6bb88f56905f.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 991.713414] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 991.713713] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2b6205ed-e5c4-406c-994c-73c1af398665 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.725126] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 991.725126] env[62600]: value = "task-1222913" [ 991.725126] env[62600]: _type = "Task" [ 991.725126] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.735570] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222913, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.837442] env[62600]: DEBUG nova.network.neutron [-] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.840784] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.841464] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.841645] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.841941] env[62600]: DEBUG oslo_concurrency.lockutils [req-9ec72425-a779-4345-bff1-51922ce6feef req-cd64fb93-e846-4f0b-91dd-845b27e18234 service nova] Acquired lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.842164] env[62600]: DEBUG nova.network.neutron [req-9ec72425-a779-4345-bff1-51922ce6feef req-cd64fb93-e846-4f0b-91dd-845b27e18234 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Refreshing network info cache for port 0bc0b585-65b4-4ebf-81be-bbc628771202 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 991.844397] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed47f76e-834e-450e-992b-f1a8344c6367 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.866850] env[62600]: DEBUG nova.virt.hardware [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 991.867125] env[62600]: DEBUG nova.virt.hardware [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 991.867292] env[62600]: DEBUG nova.virt.hardware [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 991.867480] env[62600]: DEBUG nova.virt.hardware [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 991.867632] env[62600]: DEBUG nova.virt.hardware [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 991.867787] env[62600]: DEBUG nova.virt.hardware [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 991.868025] env[62600]: DEBUG nova.virt.hardware [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 991.868184] env[62600]: DEBUG nova.virt.hardware [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 991.868356] env[62600]: DEBUG nova.virt.hardware [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 991.868526] env[62600]: DEBUG nova.virt.hardware [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 991.868849] env[62600]: DEBUG nova.virt.hardware [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 991.876226] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Reconfiguring VM to attach interface {{(pid=62600) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 991.877085] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7e140c6-60a1-43d5-9cb8-f1da180f9c2d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.896850] env[62600]: DEBUG oslo_vmware.api [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 991.896850] env[62600]: value = "task-1222914" [ 991.896850] env[62600]: _type = "Task" [ 991.896850] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.909048] env[62600]: DEBUG oslo_vmware.api [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222914, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.050912] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222912, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.237057] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222913, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.160095} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.237494] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 992.238234] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09bbe8c4-813d-48ce-8fcb-e08eed3dbd7e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.262434] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 5d7c2a17-9dee-407e-a7ee-6bb88f56905f/5d7c2a17-9dee-407e-a7ee-6bb88f56905f.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 992.262788] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5882cb6-2c42-440c-9910-0dfa9a687697 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.285200] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 992.285200] env[62600]: value = "task-1222915" [ 992.285200] env[62600]: _type = "Task" [ 992.285200] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.294574] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222915, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.341333] env[62600]: INFO nova.compute.manager [-] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Took 1.03 seconds to deallocate network for instance. [ 992.414215] env[62600]: DEBUG oslo_vmware.api [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222914, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.552810] env[62600]: DEBUG oslo_vmware.api [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222912, 'name': PowerOnVM_Task, 'duration_secs': 0.906373} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.553107] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 992.553355] env[62600]: INFO nova.compute.manager [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Took 7.85 seconds to spawn the instance on the hypervisor. [ 992.553516] env[62600]: DEBUG nova.compute.manager [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 992.554391] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4cd37b8-57b0-4b39-a47e-563fea87e9ef {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.607393] env[62600]: DEBUG nova.network.neutron [req-9ec72425-a779-4345-bff1-51922ce6feef req-cd64fb93-e846-4f0b-91dd-845b27e18234 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Updated VIF entry in instance network info cache for port 0bc0b585-65b4-4ebf-81be-bbc628771202. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 992.607918] env[62600]: DEBUG nova.network.neutron [req-9ec72425-a779-4345-bff1-51922ce6feef req-cd64fb93-e846-4f0b-91dd-845b27e18234 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Updating instance_info_cache with network_info: [{"id": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "address": "fa:16:3e:67:24:01", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8aca44-97", "ovs_interfaceid": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "24f6c3e5-a847-44bd-ba62-c4d66a2489d8", "address": "fa:16:3e:e1:5b:18", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap24f6c3e5-a8", "ovs_interfaceid": "24f6c3e5-a847-44bd-ba62-c4d66a2489d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0bc0b585-65b4-4ebf-81be-bbc628771202", "address": "fa:16:3e:da:a7:96", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bc0b585-65", "ovs_interfaceid": "0bc0b585-65b4-4ebf-81be-bbc628771202", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.797064] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222915, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.847687] env[62600]: DEBUG oslo_concurrency.lockutils [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.848139] env[62600]: DEBUG oslo_concurrency.lockutils [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.848451] env[62600]: DEBUG nova.objects.instance [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Lazy-loading 'resources' on Instance uuid e4bc0326-66d8-45e2-bcd8-11c4fab16e34 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.912672] env[62600]: DEBUG oslo_vmware.api [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222914, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.007327] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67fb5c63-e2da-445d-974f-9f5b366c17e7 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "02c6dac9-535b-436b-bbee-c8031cb5d7f5" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.007769] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67fb5c63-e2da-445d-974f-9f5b366c17e7 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "02c6dac9-535b-436b-bbee-c8031cb5d7f5" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.008064] env[62600]: DEBUG nova.compute.manager [None req-67fb5c63-e2da-445d-974f-9f5b366c17e7 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Going to confirm migration 2 {{(pid=62600) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 993.142640] env[62600]: INFO nova.compute.manager [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Took 18.45 seconds to build instance. [ 993.142640] env[62600]: DEBUG oslo_concurrency.lockutils [req-9ec72425-a779-4345-bff1-51922ce6feef req-cd64fb93-e846-4f0b-91dd-845b27e18234 service nova] Releasing lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.297153] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222915, 'name': ReconfigVM_Task, 'duration_secs': 0.886346} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.297561] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 5d7c2a17-9dee-407e-a7ee-6bb88f56905f/5d7c2a17-9dee-407e-a7ee-6bb88f56905f.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 993.298166] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-60c5dcb1-ed2f-4f46-9e5b-f5822a1c99af {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.306786] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 993.306786] env[62600]: value = "task-1222916" [ 993.306786] env[62600]: _type = "Task" [ 993.306786] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.315865] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222916, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.411820] env[62600]: DEBUG oslo_vmware.api [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222914, 'name': ReconfigVM_Task, 'duration_secs': 1.293528} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.414801] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.415163] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Reconfigured VM to attach interface {{(pid=62600) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 993.538921] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d65606a-a375-48b9-ac50-f7d4b238203d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.546803] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-731f9d72-4fb2-4424-bf80-5353feddb861 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.307290] env[62600]: DEBUG oslo_concurrency.lockutils [None req-c06f0a44-f7f8-4b99-a0ac-f517f67c4e84 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "aff91190-e19b-4f1c-8b47-9539360a4596" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.690s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.308537] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a5bb010b-27e1-4b1a-8e16-127405eab105 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "interface-88cf521d-f4a3-493a-894d-c7f799dc0443-0bc0b585-65b4-4ebf-81be-bbc628771202" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.389s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.310616] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67fb5c63-e2da-445d-974f-9f5b366c17e7 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "refresh_cache-02c6dac9-535b-436b-bbee-c8031cb5d7f5" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.310784] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67fb5c63-e2da-445d-974f-9f5b366c17e7 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "refresh_cache-02c6dac9-535b-436b-bbee-c8031cb5d7f5" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.310954] env[62600]: DEBUG nova.network.neutron [None req-67fb5c63-e2da-445d-974f-9f5b366c17e7 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 994.311147] env[62600]: DEBUG nova.objects.instance [None req-67fb5c63-e2da-445d-974f-9f5b366c17e7 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lazy-loading 'info_cache' on Instance uuid 02c6dac9-535b-436b-bbee-c8031cb5d7f5 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 994.319159] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fd1ad49-bdb0-48e1-880f-fdceb098e54f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.327655] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af3bb2c7-50d7-43d8-bf45-1278157bdce5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.331337] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222916, 'name': Rename_Task, 'duration_secs': 0.155722} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.331880] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 994.332458] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9ea93b91-4c5b-4fc5-9730-1021d3a0fccf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.345257] env[62600]: DEBUG nova.compute.provider_tree [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.345718] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 994.345718] env[62600]: value = "task-1222917" [ 994.345718] env[62600]: _type = "Task" [ 994.345718] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.354558] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222917, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.847694] env[62600]: DEBUG nova.scheduler.client.report [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 994.860992] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222917, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.103163] env[62600]: DEBUG nova.compute.manager [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Stashing vm_state: active {{(pid=62600) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 995.355950] env[62600]: DEBUG oslo_concurrency.lockutils [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.508s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.364088] env[62600]: DEBUG oslo_vmware.api [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222917, 'name': PowerOnVM_Task, 'duration_secs': 0.538306} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.364435] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 995.364733] env[62600]: INFO nova.compute.manager [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Took 8.35 seconds to spawn the instance on the hypervisor. [ 995.364977] env[62600]: DEBUG nova.compute.manager [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 995.368614] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c2821c-802b-4701-b1e4-4d568b01ad91 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.384812] env[62600]: INFO nova.scheduler.client.report [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Deleted allocations for instance e4bc0326-66d8-45e2-bcd8-11c4fab16e34 [ 995.550638] env[62600]: DEBUG nova.network.neutron [None req-67fb5c63-e2da-445d-974f-9f5b366c17e7 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Updating instance_info_cache with network_info: [{"id": "3753b6a6-f27d-4e1d-b801-91324a7e4e0e", "address": "fa:16:3e:fd:6e:a3", "network": {"id": "95818650-8227-4d11-9d14-007f06c07790", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1738152115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73e664d178f7484a9f4741b4d9450e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3753b6a6-f2", "ovs_interfaceid": "3753b6a6-f27d-4e1d-b801-91324a7e4e0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.623370] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.623655] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.783604] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "interface-88cf521d-f4a3-493a-894d-c7f799dc0443-24f6c3e5-a847-44bd-ba62-c4d66a2489d8" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.784071] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "interface-88cf521d-f4a3-493a-894d-c7f799dc0443-24f6c3e5-a847-44bd-ba62-c4d66a2489d8" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.887655] env[62600]: INFO nova.compute.manager [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Took 17.16 seconds to build instance. [ 995.894491] env[62600]: DEBUG oslo_concurrency.lockutils [None req-07a80ebb-8387-4b38-9a68-a2cc7aaf76e7 tempest-ServersAaction247Test-39236517 tempest-ServersAaction247Test-39236517-project-member] Lock "e4bc0326-66d8-45e2-bcd8-11c4fab16e34" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.777s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.053961] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67fb5c63-e2da-445d-974f-9f5b366c17e7 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "refresh_cache-02c6dac9-535b-436b-bbee-c8031cb5d7f5" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.054273] env[62600]: DEBUG nova.objects.instance [None req-67fb5c63-e2da-445d-974f-9f5b366c17e7 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lazy-loading 'migration_context' on Instance uuid 02c6dac9-535b-436b-bbee-c8031cb5d7f5 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.128312] env[62600]: INFO nova.compute.claims [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 996.290969] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.291247] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.292145] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cab7663-2caf-4b90-87ac-e3f450f25118 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.312211] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a8136e-3822-401b-8dec-5a5ecc136fc5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.342372] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Reconfiguring VM to detach interface {{(pid=62600) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 996.342657] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff89f2a9-425f-44e5-bde0-69697de5e394 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.363086] env[62600]: DEBUG oslo_vmware.api [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 996.363086] env[62600]: value = "task-1222918" [ 996.363086] env[62600]: _type = "Task" [ 996.363086] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.371359] env[62600]: DEBUG oslo_vmware.api [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222918, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.390685] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9e2d463e-7daf-496c-9c84-e10c16dd018a tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "5d7c2a17-9dee-407e-a7ee-6bb88f56905f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.665s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.557487] env[62600]: DEBUG nova.objects.base [None req-67fb5c63-e2da-445d-974f-9f5b366c17e7 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Object Instance<02c6dac9-535b-436b-bbee-c8031cb5d7f5> lazy-loaded attributes: info_cache,migration_context {{(pid=62600) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 996.558437] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75b5bd54-e37b-46c5-b8ca-ee3c5568f974 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.578612] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da80f610-7d54-483f-b81e-5b50b1b914c1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.585182] env[62600]: DEBUG oslo_vmware.api [None req-67fb5c63-e2da-445d-974f-9f5b366c17e7 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 996.585182] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b82c78-70c8-fae7-8a6f-68bd04353c85" [ 996.585182] env[62600]: _type = "Task" [ 996.585182] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.594152] env[62600]: DEBUG oslo_vmware.api [None req-67fb5c63-e2da-445d-974f-9f5b366c17e7 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b82c78-70c8-fae7-8a6f-68bd04353c85, 'name': SearchDatastore_Task} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.595491] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67fb5c63-e2da-445d-974f-9f5b366c17e7 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.634943] env[62600]: INFO nova.compute.resource_tracker [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Updating resource usage from migration e8e97f1c-3473-4838-bfb2-5e02bd200b70 [ 996.810741] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4ed12d2-4ce4-4c48-ae10-094721a277e9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.819897] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f79e19ce-7ee2-4375-a5d0-7075c8a3863d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.856783] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a3ceadb-8671-4972-93c3-a50d02ae0cba {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.873464] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22f73a05-be67-456e-8b34-1dc08cf19044 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.884696] env[62600]: DEBUG oslo_vmware.api [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222918, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.894282] env[62600]: DEBUG nova.compute.provider_tree [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.376593] env[62600]: DEBUG oslo_vmware.api [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222918, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.397911] env[62600]: DEBUG nova.scheduler.client.report [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 997.543324] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "a20222a7-f29b-468f-8105-50ff9e31bab2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.543570] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "a20222a7-f29b-468f-8105-50ff9e31bab2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.876021] env[62600]: DEBUG oslo_vmware.api [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222918, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.904044] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.280s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.904272] env[62600]: INFO nova.compute.manager [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Migrating [ 997.910857] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67fb5c63-e2da-445d-974f-9f5b366c17e7 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.316s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.045850] env[62600]: DEBUG nova.compute.manager [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 998.376523] env[62600]: DEBUG oslo_vmware.api [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222918, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.421714] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "refresh_cache-aff91190-e19b-4f1c-8b47-9539360a4596" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.422067] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired lock "refresh_cache-aff91190-e19b-4f1c-8b47-9539360a4596" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.422067] env[62600]: DEBUG nova.network.neutron [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 998.565892] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.596892] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30b564e-8cb4-431c-b81c-0e241b39f645 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.605050] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22434ec5-9aad-49ee-a07e-2c1cc2c13235 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.634748] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b8abbc1-013e-4821-a4cd-6c296ad4aa67 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.642648] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ab94c1d-c25d-479e-a96f-6b139aae1ec7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.656744] env[62600]: DEBUG nova.compute.provider_tree [None req-67fb5c63-e2da-445d-974f-9f5b366c17e7 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.877457] env[62600]: DEBUG oslo_vmware.api [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222918, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.136213] env[62600]: DEBUG nova.network.neutron [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Updating instance_info_cache with network_info: [{"id": "7af8f843-5635-4c50-bc15-f8356b10e6f7", "address": "fa:16:3e:fc:4b:8b", "network": {"id": "5ef67e8a-5613-4cdb-9f35-0a18720657ff", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1644754500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bab74e3ae78248909dbcd483abbb8da2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7af8f843-56", "ovs_interfaceid": "7af8f843-5635-4c50-bc15-f8356b10e6f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.159440] env[62600]: DEBUG nova.scheduler.client.report [None req-67fb5c63-e2da-445d-974f-9f5b366c17e7 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 999.378925] env[62600]: DEBUG oslo_vmware.api [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222918, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.639160] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Releasing lock "refresh_cache-aff91190-e19b-4f1c-8b47-9539360a4596" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.879045] env[62600]: DEBUG oslo_vmware.api [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222918, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.169407] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67fb5c63-e2da-445d-974f-9f5b366c17e7 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.258s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.173346] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.607s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.179380] env[62600]: INFO nova.compute.claims [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1000.379946] env[62600]: DEBUG oslo_vmware.api [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222918, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.738859] env[62600]: INFO nova.scheduler.client.report [None req-67fb5c63-e2da-445d-974f-9f5b366c17e7 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleted allocation for migration 8d35c436-258f-4615-a3a2-eb29f7f59091 [ 1000.879912] env[62600]: DEBUG oslo_vmware.api [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222918, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.154244] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18142911-e49e-4cee-9685-186adb381e69 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.174870] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Updating instance 'aff91190-e19b-4f1c-8b47-9539360a4596' progress to 0 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1001.249016] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67fb5c63-e2da-445d-974f-9f5b366c17e7 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "02c6dac9-535b-436b-bbee-c8031cb5d7f5" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.241s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.372067] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-578a92b4-7b65-4c32-aaaf-00c7467b6a34 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.385073] env[62600]: DEBUG oslo_vmware.api [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222918, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.387737] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-588e5440-2668-4534-ab61-c50d50a64b51 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.417790] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83aeaf64-c374-49c5-9899-0bdf74867a11 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.425310] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f0cea1f-18a0-4314-9f64-87ae9b030fc9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.438707] env[62600]: DEBUG nova.compute.provider_tree [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.681525] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1001.681698] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-344a6812-167c-4196-8aa4-a7bbccb7740c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.690916] env[62600]: DEBUG oslo_vmware.api [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 1001.690916] env[62600]: value = "task-1222919" [ 1001.690916] env[62600]: _type = "Task" [ 1001.690916] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.701214] env[62600]: DEBUG oslo_vmware.api [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222919, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.880907] env[62600]: DEBUG oslo_vmware.api [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222918, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.942215] env[62600]: DEBUG nova.scheduler.client.report [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1002.202811] env[62600]: DEBUG oslo_vmware.api [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222919, 'name': PowerOffVM_Task, 'duration_secs': 0.224181} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.203107] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1002.203295] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Updating instance 'aff91190-e19b-4f1c-8b47-9539360a4596' progress to 17 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1002.223732] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "02c6dac9-535b-436b-bbee-c8031cb5d7f5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.223985] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "02c6dac9-535b-436b-bbee-c8031cb5d7f5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.224210] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "02c6dac9-535b-436b-bbee-c8031cb5d7f5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.224396] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "02c6dac9-535b-436b-bbee-c8031cb5d7f5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.224569] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "02c6dac9-535b-436b-bbee-c8031cb5d7f5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.226749] env[62600]: INFO nova.compute.manager [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Terminating instance [ 1002.228457] env[62600]: DEBUG nova.compute.manager [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1002.228654] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1002.229508] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf01441-9f05-4318-b445-a23553e90887 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.237863] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1002.238128] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee1adc83-330e-437f-a592-bc994e67435e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.246149] env[62600]: DEBUG oslo_vmware.api [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1002.246149] env[62600]: value = "task-1222920" [ 1002.246149] env[62600]: _type = "Task" [ 1002.246149] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.254592] env[62600]: DEBUG oslo_vmware.api [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222920, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.381844] env[62600]: DEBUG oslo_vmware.api [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222918, 'name': ReconfigVM_Task, 'duration_secs': 5.750884} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.381844] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.381844] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Reconfigured VM to detach interface {{(pid=62600) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1002.447580] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.275s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.448148] env[62600]: DEBUG nova.compute.manager [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1002.637174] env[62600]: DEBUG nova.compute.manager [req-5eaeb784-1c5d-4671-b75b-f3d81fcbb138 req-db5599b5-9b95-4afa-9890-32c4bf03f58b service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Received event network-vif-deleted-24f6c3e5-a847-44bd-ba62-c4d66a2489d8 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.637368] env[62600]: INFO nova.compute.manager [req-5eaeb784-1c5d-4671-b75b-f3d81fcbb138 req-db5599b5-9b95-4afa-9890-32c4bf03f58b service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Neutron deleted interface 24f6c3e5-a847-44bd-ba62-c4d66a2489d8; detaching it from the instance and deleting it from the info cache [ 1002.637612] env[62600]: DEBUG nova.network.neutron [req-5eaeb784-1c5d-4671-b75b-f3d81fcbb138 req-db5599b5-9b95-4afa-9890-32c4bf03f58b service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Updating instance_info_cache with network_info: [{"id": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "address": "fa:16:3e:67:24:01", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8aca44-97", "ovs_interfaceid": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0bc0b585-65b4-4ebf-81be-bbc628771202", "address": "fa:16:3e:da:a7:96", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bc0b585-65", "ovs_interfaceid": "0bc0b585-65b4-4ebf-81be-bbc628771202", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.710685] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1002.711012] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1002.711268] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1002.711544] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1002.711724] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1002.711960] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1002.712263] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1002.712513] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1002.712859] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1002.713135] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1002.713411] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1002.719356] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ee9e1ee-de42-45c0-86ad-56bfb0caed5f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.740741] env[62600]: DEBUG oslo_vmware.api [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 1002.740741] env[62600]: value = "task-1222921" [ 1002.740741] env[62600]: _type = "Task" [ 1002.740741] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.750787] env[62600]: DEBUG oslo_vmware.api [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222921, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.762416] env[62600]: DEBUG oslo_vmware.api [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222920, 'name': PowerOffVM_Task, 'duration_secs': 0.156624} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.762747] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1002.762977] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1002.763270] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-892afdfc-ab31-47ae-a231-eb458d037744 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.787763] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "0416ef03-fd45-405f-9047-76b34a839217" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.788014] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "0416ef03-fd45-405f-9047-76b34a839217" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.882884] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1002.883360] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1002.883509] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleting the datastore file [datastore2] 02c6dac9-535b-436b-bbee-c8031cb5d7f5 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1002.883794] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5edbca00-60f8-415f-9110-298271bcadb7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.891560] env[62600]: DEBUG oslo_vmware.api [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1002.891560] env[62600]: value = "task-1222923" [ 1002.891560] env[62600]: _type = "Task" [ 1002.891560] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.901352] env[62600]: DEBUG oslo_vmware.api [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222923, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.953334] env[62600]: DEBUG nova.compute.utils [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1002.954888] env[62600]: DEBUG nova.compute.manager [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1002.955086] env[62600]: DEBUG nova.network.neutron [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1003.003203] env[62600]: DEBUG nova.policy [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3e59998217a4b18a6f3f01142a5e440', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f91091f83ee4a2091507ca994e3d52f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 1003.140699] env[62600]: DEBUG oslo_concurrency.lockutils [req-5eaeb784-1c5d-4671-b75b-f3d81fcbb138 req-db5599b5-9b95-4afa-9890-32c4bf03f58b service nova] Acquiring lock "88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.140699] env[62600]: DEBUG oslo_concurrency.lockutils [req-5eaeb784-1c5d-4671-b75b-f3d81fcbb138 req-db5599b5-9b95-4afa-9890-32c4bf03f58b service nova] Acquired lock "88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.141278] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77830cb3-0dda-40dd-8265-4d2b25c23d01 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.160918] env[62600]: DEBUG oslo_concurrency.lockutils [req-5eaeb784-1c5d-4671-b75b-f3d81fcbb138 req-db5599b5-9b95-4afa-9890-32c4bf03f58b service nova] Releasing lock "88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.161201] env[62600]: WARNING nova.compute.manager [req-5eaeb784-1c5d-4671-b75b-f3d81fcbb138 req-db5599b5-9b95-4afa-9890-32c4bf03f58b service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Detach interface failed, port_id=24f6c3e5-a847-44bd-ba62-c4d66a2489d8, reason: No device with interface-id 24f6c3e5-a847-44bd-ba62-c4d66a2489d8 exists on VM: nova.exception.NotFound: No device with interface-id 24f6c3e5-a847-44bd-ba62-c4d66a2489d8 exists on VM [ 1003.251290] env[62600]: DEBUG oslo_vmware.api [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222921, 'name': ReconfigVM_Task, 'duration_secs': 0.181494} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.251618] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Updating instance 'aff91190-e19b-4f1c-8b47-9539360a4596' progress to 33 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1003.255548] env[62600]: DEBUG nova.network.neutron [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Successfully created port: 0ffa19db-f509-4c53-be0e-7a1692d2f4b7 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1003.290697] env[62600]: DEBUG nova.compute.manager [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1003.328717] env[62600]: DEBUG nova.compute.manager [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Received event network-vif-deleted-0bc0b585-65b4-4ebf-81be-bbc628771202 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1003.328717] env[62600]: INFO nova.compute.manager [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Neutron deleted interface 0bc0b585-65b4-4ebf-81be-bbc628771202; detaching it from the instance and deleting it from the info cache [ 1003.328932] env[62600]: DEBUG nova.network.neutron [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Updating instance_info_cache with network_info: [{"id": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "address": "fa:16:3e:67:24:01", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8aca44-97", "ovs_interfaceid": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.402161] env[62600]: DEBUG oslo_vmware.api [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222923, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.201323} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.402436] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1003.402627] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1003.402813] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1003.402994] env[62600]: INFO nova.compute.manager [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1003.403261] env[62600]: DEBUG oslo.service.loopingcall [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1003.403457] env[62600]: DEBUG nova.compute.manager [-] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1003.403670] env[62600]: DEBUG nova.network.neutron [-] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1003.458551] env[62600]: DEBUG nova.compute.manager [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1003.611823] env[62600]: DEBUG oslo_concurrency.lockutils [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "39deb498-6bf0-4f3b-932b-8068fc48271e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.611823] env[62600]: DEBUG oslo_concurrency.lockutils [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "39deb498-6bf0-4f3b-932b-8068fc48271e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.611990] env[62600]: DEBUG oslo_concurrency.lockutils [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "39deb498-6bf0-4f3b-932b-8068fc48271e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.612868] env[62600]: DEBUG oslo_concurrency.lockutils [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "39deb498-6bf0-4f3b-932b-8068fc48271e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.612868] env[62600]: DEBUG oslo_concurrency.lockutils [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "39deb498-6bf0-4f3b-932b-8068fc48271e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.614890] env[62600]: INFO nova.compute.manager [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Terminating instance [ 1003.617246] env[62600]: DEBUG nova.compute.manager [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1003.617595] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1003.618144] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.618444] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.621021] env[62600]: DEBUG nova.network.neutron [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1003.621021] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfdcfcf6-bd3e-4186-82d5-d9ac8b6c2612 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.629487] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1003.630238] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ed8224d2-3e72-40a3-8455-01b3a6782f7c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.639356] env[62600]: DEBUG oslo_vmware.api [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1003.639356] env[62600]: value = "task-1222924" [ 1003.639356] env[62600]: _type = "Task" [ 1003.639356] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.649289] env[62600]: DEBUG oslo_vmware.api [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222924, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.760918] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1003.761291] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1003.761506] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1003.761765] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1003.761958] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1003.762132] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1003.762352] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1003.762521] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1003.762757] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1003.762953] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1003.763158] env[62600]: DEBUG nova.virt.hardware [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1003.768606] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Reconfiguring VM instance instance-0000005d to detach disk 2000 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1003.768896] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2260843b-e71c-49c5-88c3-b2b3c3cabb26 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.789520] env[62600]: DEBUG oslo_vmware.api [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 1003.789520] env[62600]: value = "task-1222925" [ 1003.789520] env[62600]: _type = "Task" [ 1003.789520] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.803167] env[62600]: DEBUG oslo_vmware.api [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222925, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.814646] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.815024] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.816872] env[62600]: INFO nova.compute.claims [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1003.834877] env[62600]: DEBUG oslo_concurrency.lockutils [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] Acquiring lock "88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.835124] env[62600]: DEBUG oslo_concurrency.lockutils [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] Acquired lock "88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.835564] env[62600]: DEBUG oslo_concurrency.lockutils [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "88cf521d-f4a3-493a-894d-c7f799dc0443" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.836494] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1059e955-8ece-4001-aa0c-d90befbdd2be {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.856805] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05459ccb-a4a1-44dc-b137-061ea79a3a73 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.885158] env[62600]: DEBUG nova.virt.vmwareapi.vmops [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Reconfiguring VM to detach interface {{(pid=62600) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1003.885796] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2536bb1a-1251-405d-b164-2680d2228c93 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.906468] env[62600]: DEBUG oslo_vmware.api [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] Waiting for the task: (returnval){ [ 1003.906468] env[62600]: value = "task-1222926" [ 1003.906468] env[62600]: _type = "Task" [ 1003.906468] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.915029] env[62600]: DEBUG oslo_vmware.api [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] Task: {'id': task-1222926, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.150755] env[62600]: DEBUG oslo_vmware.api [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222924, 'name': PowerOffVM_Task, 'duration_secs': 0.249928} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.150755] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1004.150950] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1004.151057] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-87597364-b5b3-40af-9781-3483d9103087 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.199074] env[62600]: DEBUG nova.network.neutron [-] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.246858] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1004.247129] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1004.247328] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Deleting the datastore file [datastore1] 39deb498-6bf0-4f3b-932b-8068fc48271e {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1004.247596] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-345b928c-2820-44af-a7f6-60f3312272d9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.255490] env[62600]: DEBUG oslo_vmware.api [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1004.255490] env[62600]: value = "task-1222928" [ 1004.255490] env[62600]: _type = "Task" [ 1004.255490] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.263536] env[62600]: DEBUG oslo_vmware.api [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222928, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.301166] env[62600]: DEBUG oslo_vmware.api [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222925, 'name': ReconfigVM_Task, 'duration_secs': 0.175693} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.301463] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Reconfigured VM instance instance-0000005d to detach disk 2000 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1004.302284] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce268297-772d-476c-a101-9a01ac72bb35 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.326724] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] aff91190-e19b-4f1c-8b47-9539360a4596/aff91190-e19b-4f1c-8b47-9539360a4596.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1004.329243] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a3f98d0-17ff-415c-b406-8d2bb54d6f46 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.349758] env[62600]: DEBUG oslo_vmware.api [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 1004.349758] env[62600]: value = "task-1222929" [ 1004.349758] env[62600]: _type = "Task" [ 1004.349758] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.358951] env[62600]: DEBUG oslo_vmware.api [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222929, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.417517] env[62600]: DEBUG oslo_vmware.api [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] Task: {'id': task-1222926, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.456136] env[62600]: DEBUG nova.network.neutron [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Updating instance_info_cache with network_info: [{"id": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "address": "fa:16:3e:67:24:01", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a8aca44-97", "ovs_interfaceid": "3a8aca44-9791-4493-a1fd-114c9f20fb3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.467987] env[62600]: DEBUG nova.compute.manager [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1004.495318] env[62600]: DEBUG nova.virt.hardware [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1004.495842] env[62600]: DEBUG nova.virt.hardware [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1004.496221] env[62600]: DEBUG nova.virt.hardware [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1004.497134] env[62600]: DEBUG nova.virt.hardware [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1004.497134] env[62600]: DEBUG nova.virt.hardware [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1004.497134] env[62600]: DEBUG nova.virt.hardware [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1004.497327] env[62600]: DEBUG nova.virt.hardware [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1004.497578] env[62600]: DEBUG nova.virt.hardware [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1004.497833] env[62600]: DEBUG nova.virt.hardware [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1004.498137] env[62600]: DEBUG nova.virt.hardware [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1004.498424] env[62600]: DEBUG nova.virt.hardware [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1004.499319] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d1b5de3-9d7a-47a7-a510-8b8a9bd2cac8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.508746] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ede9ef8-eabd-4bf6-b91e-9e68cf5f38b9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.702268] env[62600]: INFO nova.compute.manager [-] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Took 1.30 seconds to deallocate network for instance. [ 1004.766080] env[62600]: DEBUG oslo_vmware.api [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222928, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148501} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.766343] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1004.766551] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1004.766731] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1004.766953] env[62600]: INFO nova.compute.manager [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1004.767449] env[62600]: DEBUG oslo.service.loopingcall [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.767449] env[62600]: DEBUG nova.compute.manager [-] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1004.767449] env[62600]: DEBUG nova.network.neutron [-] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1004.841087] env[62600]: DEBUG nova.compute.manager [req-c2478fd3-96d2-467d-a981-8602f14a29c2 req-1cbca97d-fb6f-4003-8ea5-4912e9e94342 service nova] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Received event network-vif-plugged-0ffa19db-f509-4c53-be0e-7a1692d2f4b7 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.841329] env[62600]: DEBUG oslo_concurrency.lockutils [req-c2478fd3-96d2-467d-a981-8602f14a29c2 req-1cbca97d-fb6f-4003-8ea5-4912e9e94342 service nova] Acquiring lock "a20222a7-f29b-468f-8105-50ff9e31bab2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.841548] env[62600]: DEBUG oslo_concurrency.lockutils [req-c2478fd3-96d2-467d-a981-8602f14a29c2 req-1cbca97d-fb6f-4003-8ea5-4912e9e94342 service nova] Lock "a20222a7-f29b-468f-8105-50ff9e31bab2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.841726] env[62600]: DEBUG oslo_concurrency.lockutils [req-c2478fd3-96d2-467d-a981-8602f14a29c2 req-1cbca97d-fb6f-4003-8ea5-4912e9e94342 service nova] Lock "a20222a7-f29b-468f-8105-50ff9e31bab2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.841906] env[62600]: DEBUG nova.compute.manager [req-c2478fd3-96d2-467d-a981-8602f14a29c2 req-1cbca97d-fb6f-4003-8ea5-4912e9e94342 service nova] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] No waiting events found dispatching network-vif-plugged-0ffa19db-f509-4c53-be0e-7a1692d2f4b7 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1004.842414] env[62600]: WARNING nova.compute.manager [req-c2478fd3-96d2-467d-a981-8602f14a29c2 req-1cbca97d-fb6f-4003-8ea5-4912e9e94342 service nova] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Received unexpected event network-vif-plugged-0ffa19db-f509-4c53-be0e-7a1692d2f4b7 for instance with vm_state building and task_state spawning. [ 1004.860328] env[62600]: DEBUG oslo_vmware.api [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222929, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.922394] env[62600]: DEBUG oslo_vmware.api [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] Task: {'id': task-1222926, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.936242] env[62600]: DEBUG nova.network.neutron [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Successfully updated port: 0ffa19db-f509-4c53-be0e-7a1692d2f4b7 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1004.958901] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "refresh_cache-88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.036894] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a0d5dc1-8b5c-43dd-a164-d58f624babf9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.046922] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ed07a4-07f1-42d8-8bf2-3308a5a04440 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.076841] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2fda7f-4f7e-44cf-9924-65c0a569560d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.084197] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be1656c5-265b-45b0-a176-df0caa62815b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.097293] env[62600]: DEBUG nova.compute.provider_tree [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1005.209456] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.349277] env[62600]: DEBUG nova.compute.manager [req-edc51deb-d68c-46ba-881f-73aec53a1191 req-0c4c4dcb-e9de-4394-9922-74dfe9e45ade service nova] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Received event network-vif-deleted-3753b6a6-f27d-4e1d-b801-91324a7e4e0e {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1005.361310] env[62600]: DEBUG oslo_vmware.api [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222929, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.418763] env[62600]: DEBUG oslo_vmware.api [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] Task: {'id': task-1222926, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.440748] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "refresh_cache-a20222a7-f29b-468f-8105-50ff9e31bab2" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.440748] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "refresh_cache-a20222a7-f29b-468f-8105-50ff9e31bab2" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.440748] env[62600]: DEBUG nova.network.neutron [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1005.465319] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3786c6fc-d7fc-48d9-a6b2-2f4b57d3d4d7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "interface-88cf521d-f4a3-493a-894d-c7f799dc0443-24f6c3e5-a847-44bd-ba62-c4d66a2489d8" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.681s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.618396] env[62600]: ERROR nova.scheduler.client.report [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [req-b629c7a6-548a-47e7-81b1-61cf1fd6425a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 664af347-7147-4bf5-9019-9ae15cb4aa82. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b629c7a6-548a-47e7-81b1-61cf1fd6425a"}]} [ 1005.636258] env[62600]: DEBUG nova.scheduler.client.report [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Refreshing inventories for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1005.651831] env[62600]: DEBUG nova.scheduler.client.report [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Updating ProviderTree inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1005.652083] env[62600]: DEBUG nova.compute.provider_tree [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1005.664763] env[62600]: DEBUG nova.scheduler.client.report [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Refreshing aggregate associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, aggregates: None {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1005.681833] env[62600]: DEBUG nova.scheduler.client.report [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Refreshing trait associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1005.842053] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72f260c2-0c23-44f7-9062-2d513d877b9f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.850108] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5e80eb9-5bf5-4a3e-bf30-0e4da7705e32 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.861228] env[62600]: DEBUG oslo_vmware.api [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222929, 'name': ReconfigVM_Task, 'duration_secs': 1.280632} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.887554] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Reconfigured VM instance instance-0000005d to attach disk [datastore1] aff91190-e19b-4f1c-8b47-9539360a4596/aff91190-e19b-4f1c-8b47-9539360a4596.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1005.887923] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Updating instance 'aff91190-e19b-4f1c-8b47-9539360a4596' progress to 50 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1005.891627] env[62600]: DEBUG nova.network.neutron [-] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.894345] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24eba0f3-2d34-4412-8922-563ca44dcb27 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.903239] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24cfa97d-2abf-41fc-a583-3942c7f0981c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.922256] env[62600]: DEBUG nova.compute.provider_tree [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1005.929478] env[62600]: DEBUG oslo_vmware.api [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] Task: {'id': task-1222926, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.973925] env[62600]: DEBUG nova.network.neutron [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1006.098215] env[62600]: DEBUG nova.network.neutron [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Updating instance_info_cache with network_info: [{"id": "0ffa19db-f509-4c53-be0e-7a1692d2f4b7", "address": "fa:16:3e:61:30:b6", "network": {"id": "1d2052cf-5be1-48e7-887a-18f367b7823f", "bridge": "br-int", "label": "tempest-ServersTestJSON-106099679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f91091f83ee4a2091507ca994e3d52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ffa19db-f5", "ovs_interfaceid": "0ffa19db-f509-4c53-be0e-7a1692d2f4b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.395566] env[62600]: INFO nova.compute.manager [-] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Took 1.63 seconds to deallocate network for instance. [ 1006.396436] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-096f90a9-0016-47a8-8532-67e913d61934 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.421920] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56a40e90-573b-4709-8443-9ce825db12b2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.432366] env[62600]: DEBUG oslo_vmware.api [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] Task: {'id': task-1222926, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.445716] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Updating instance 'aff91190-e19b-4f1c-8b47-9539360a4596' progress to 67 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1006.451726] env[62600]: DEBUG nova.scheduler.client.report [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 129 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1006.451968] env[62600]: DEBUG nova.compute.provider_tree [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 129 to 130 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1006.452171] env[62600]: DEBUG nova.compute.provider_tree [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1006.601013] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "refresh_cache-a20222a7-f29b-468f-8105-50ff9e31bab2" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.601374] env[62600]: DEBUG nova.compute.manager [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Instance network_info: |[{"id": "0ffa19db-f509-4c53-be0e-7a1692d2f4b7", "address": "fa:16:3e:61:30:b6", "network": {"id": "1d2052cf-5be1-48e7-887a-18f367b7823f", "bridge": "br-int", "label": "tempest-ServersTestJSON-106099679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f91091f83ee4a2091507ca994e3d52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ffa19db-f5", "ovs_interfaceid": "0ffa19db-f509-4c53-be0e-7a1692d2f4b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1006.601833] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:61:30:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '028bae2d-fe6c-4207-b4a3-3fab45fbf1d6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0ffa19db-f509-4c53-be0e-7a1692d2f4b7', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1006.609302] env[62600]: DEBUG oslo.service.loopingcall [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1006.609536] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1006.609759] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-db9aedb8-13f7-451a-aac4-bcb2ea3625ff {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.631956] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1006.631956] env[62600]: value = "task-1222930" [ 1006.631956] env[62600]: _type = "Task" [ 1006.631956] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.640446] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222930, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.870156] env[62600]: DEBUG nova.compute.manager [req-7cfeef58-917d-4926-8acb-e84d3f5c9983 req-24d13fa4-bd58-41b1-abe4-9933d49b6247 service nova] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Received event network-changed-0ffa19db-f509-4c53-be0e-7a1692d2f4b7 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1006.870362] env[62600]: DEBUG nova.compute.manager [req-7cfeef58-917d-4926-8acb-e84d3f5c9983 req-24d13fa4-bd58-41b1-abe4-9933d49b6247 service nova] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Refreshing instance network info cache due to event network-changed-0ffa19db-f509-4c53-be0e-7a1692d2f4b7. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1006.870598] env[62600]: DEBUG oslo_concurrency.lockutils [req-7cfeef58-917d-4926-8acb-e84d3f5c9983 req-24d13fa4-bd58-41b1-abe4-9933d49b6247 service nova] Acquiring lock "refresh_cache-a20222a7-f29b-468f-8105-50ff9e31bab2" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.870748] env[62600]: DEBUG oslo_concurrency.lockutils [req-7cfeef58-917d-4926-8acb-e84d3f5c9983 req-24d13fa4-bd58-41b1-abe4-9933d49b6247 service nova] Acquired lock "refresh_cache-a20222a7-f29b-468f-8105-50ff9e31bab2" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.870915] env[62600]: DEBUG nova.network.neutron [req-7cfeef58-917d-4926-8acb-e84d3f5c9983 req-24d13fa4-bd58-41b1-abe4-9933d49b6247 service nova] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Refreshing network info cache for port 0ffa19db-f509-4c53-be0e-7a1692d2f4b7 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1006.922943] env[62600]: DEBUG oslo_concurrency.lockutils [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.929385] env[62600]: DEBUG oslo_vmware.api [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] Task: {'id': task-1222926, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.956067] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.141s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.956587] env[62600]: DEBUG nova.compute.manager [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1006.959162] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.750s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.959377] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.961989] env[62600]: DEBUG oslo_concurrency.lockutils [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.039s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.962274] env[62600]: DEBUG nova.objects.instance [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lazy-loading 'resources' on Instance uuid 39deb498-6bf0-4f3b-932b-8068fc48271e {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1006.984036] env[62600]: INFO nova.scheduler.client.report [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleted allocations for instance 02c6dac9-535b-436b-bbee-c8031cb5d7f5 [ 1007.038859] env[62600]: DEBUG nova.network.neutron [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Port 7af8f843-5635-4c50-bc15-f8356b10e6f7 binding to destination host cpu-1 is already ACTIVE {{(pid=62600) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1007.144279] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222930, 'name': CreateVM_Task, 'duration_secs': 0.384049} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.144985] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1007.145370] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.145545] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.145870] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1007.146157] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77e570e3-6c30-47bf-a7b7-114e782ca37b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.151656] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1007.151656] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52a4816b-0163-06b1-4779-6cc6181660d1" [ 1007.151656] env[62600]: _type = "Task" [ 1007.151656] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.159727] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52a4816b-0163-06b1-4779-6cc6181660d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.376630] env[62600]: DEBUG nova.compute.manager [req-0be74b0d-30af-4add-a3a9-dbb4c9d3a728 req-fab8a04c-823e-4e2a-bff4-ff5d63069626 service nova] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Received event network-vif-deleted-ead68533-9f71-4ab4-a47d-af3e0c397bd4 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1007.427973] env[62600]: DEBUG oslo_vmware.api [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] Task: {'id': task-1222926, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.465086] env[62600]: DEBUG nova.compute.utils [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1007.468955] env[62600]: DEBUG nova.compute.manager [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1007.469073] env[62600]: DEBUG nova.network.neutron [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1007.491685] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67e3a7f0-152a-49d5-99c2-b6f5f9bae784 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "02c6dac9-535b-436b-bbee-c8031cb5d7f5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.267s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.512345] env[62600]: DEBUG nova.policy [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bd0d91fa2ac74c38b9e0ebf39c3baa7a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73e664d178f7484a9f4741b4d9450e68', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 1007.571957] env[62600]: DEBUG nova.network.neutron [req-7cfeef58-917d-4926-8acb-e84d3f5c9983 req-24d13fa4-bd58-41b1-abe4-9933d49b6247 service nova] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Updated VIF entry in instance network info cache for port 0ffa19db-f509-4c53-be0e-7a1692d2f4b7. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1007.572354] env[62600]: DEBUG nova.network.neutron [req-7cfeef58-917d-4926-8acb-e84d3f5c9983 req-24d13fa4-bd58-41b1-abe4-9933d49b6247 service nova] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Updating instance_info_cache with network_info: [{"id": "0ffa19db-f509-4c53-be0e-7a1692d2f4b7", "address": "fa:16:3e:61:30:b6", "network": {"id": "1d2052cf-5be1-48e7-887a-18f367b7823f", "bridge": "br-int", "label": "tempest-ServersTestJSON-106099679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f91091f83ee4a2091507ca994e3d52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ffa19db-f5", "ovs_interfaceid": "0ffa19db-f509-4c53-be0e-7a1692d2f4b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.631773] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db37a785-a690-44df-8678-1f908c29251c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.640709] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15bbdd2d-0da2-4ca5-a1ad-13a73ea7eec0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.674333] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-278747a7-c3d4-4c9f-9ea0-39f517115b8f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.686050] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fcf4c09-2d3b-46f7-8359-261749506a65 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.689719] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52a4816b-0163-06b1-4779-6cc6181660d1, 'name': SearchDatastore_Task, 'duration_secs': 0.010774} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.690025] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.690266] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1007.690500] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.690652] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.690836] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1007.691426] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cd440f9d-6e9f-4703-ada9-b287c51544fd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.702101] env[62600]: DEBUG nova.compute.provider_tree [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.710506] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1007.710696] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1007.712526] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2cc38d9f-0a04-4e53-a927-22de8f372bed {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.718452] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1007.718452] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528032fe-10c1-82e5-0eaf-9cdca02caaf5" [ 1007.718452] env[62600]: _type = "Task" [ 1007.718452] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.729251] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528032fe-10c1-82e5-0eaf-9cdca02caaf5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.787426] env[62600]: DEBUG nova.network.neutron [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Successfully created port: bb15b28d-be8f-4fff-b8a6-0801d3e0541a {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1007.928579] env[62600]: DEBUG oslo_vmware.api [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] Task: {'id': task-1222926, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.969545] env[62600]: DEBUG nova.compute.manager [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1008.061714] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "aff91190-e19b-4f1c-8b47-9539360a4596-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.061714] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "aff91190-e19b-4f1c-8b47-9539360a4596-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.061893] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "aff91190-e19b-4f1c-8b47-9539360a4596-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.078412] env[62600]: DEBUG oslo_concurrency.lockutils [req-7cfeef58-917d-4926-8acb-e84d3f5c9983 req-24d13fa4-bd58-41b1-abe4-9933d49b6247 service nova] Releasing lock "refresh_cache-a20222a7-f29b-468f-8105-50ff9e31bab2" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.206267] env[62600]: DEBUG nova.scheduler.client.report [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1008.231049] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528032fe-10c1-82e5-0eaf-9cdca02caaf5, 'name': SearchDatastore_Task, 'duration_secs': 0.008926} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.231049] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77cf6ae5-3976-45d1-8deb-c32a3ac793ba {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.236362] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1008.236362] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]529e759b-ad73-a1c9-0778-1f935a0ad1ce" [ 1008.236362] env[62600]: _type = "Task" [ 1008.236362] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.244304] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]529e759b-ad73-a1c9-0778-1f935a0ad1ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.429738] env[62600]: DEBUG oslo_vmware.api [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] Task: {'id': task-1222926, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.711280] env[62600]: DEBUG oslo_concurrency.lockutils [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.749s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.731262] env[62600]: INFO nova.scheduler.client.report [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Deleted allocations for instance 39deb498-6bf0-4f3b-932b-8068fc48271e [ 1008.751036] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]529e759b-ad73-a1c9-0778-1f935a0ad1ce, 'name': SearchDatastore_Task, 'duration_secs': 0.00987} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.751262] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.751673] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] a20222a7-f29b-468f-8105-50ff9e31bab2/a20222a7-f29b-468f-8105-50ff9e31bab2.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1008.752135] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cdfde6ec-0678-4dc9-859f-469228e3632e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.761456] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1008.761456] env[62600]: value = "task-1222931" [ 1008.761456] env[62600]: _type = "Task" [ 1008.761456] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.771187] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222931, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.934063] env[62600]: DEBUG oslo_vmware.api [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] Task: {'id': task-1222926, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.978712] env[62600]: DEBUG nova.compute.manager [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1009.006594] env[62600]: DEBUG nova.virt.hardware [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1009.006861] env[62600]: DEBUG nova.virt.hardware [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1009.007035] env[62600]: DEBUG nova.virt.hardware [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1009.007291] env[62600]: DEBUG nova.virt.hardware [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1009.007480] env[62600]: DEBUG nova.virt.hardware [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1009.007640] env[62600]: DEBUG nova.virt.hardware [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1009.007859] env[62600]: DEBUG nova.virt.hardware [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1009.008056] env[62600]: DEBUG nova.virt.hardware [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1009.008249] env[62600]: DEBUG nova.virt.hardware [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1009.008421] env[62600]: DEBUG nova.virt.hardware [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1009.008635] env[62600]: DEBUG nova.virt.hardware [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1009.009717] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf70bdf3-321a-47dc-b141-3ee5b3f06c95 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.019968] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-530c4827-f228-41a2-923d-4a52a53c866c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.100346] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "refresh_cache-aff91190-e19b-4f1c-8b47-9539360a4596" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.100543] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired lock "refresh_cache-aff91190-e19b-4f1c-8b47-9539360a4596" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.100774] env[62600]: DEBUG nova.network.neutron [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1009.190164] env[62600]: DEBUG nova.compute.manager [req-5103e08b-bfa9-4783-97ab-5a49959739a1 req-af2ee6dc-7ed1-4d48-a379-9c2e8e2bcc70 service nova] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Received event network-vif-plugged-bb15b28d-be8f-4fff-b8a6-0801d3e0541a {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1009.190405] env[62600]: DEBUG oslo_concurrency.lockutils [req-5103e08b-bfa9-4783-97ab-5a49959739a1 req-af2ee6dc-7ed1-4d48-a379-9c2e8e2bcc70 service nova] Acquiring lock "0416ef03-fd45-405f-9047-76b34a839217-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.190626] env[62600]: DEBUG oslo_concurrency.lockutils [req-5103e08b-bfa9-4783-97ab-5a49959739a1 req-af2ee6dc-7ed1-4d48-a379-9c2e8e2bcc70 service nova] Lock "0416ef03-fd45-405f-9047-76b34a839217-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.190792] env[62600]: DEBUG oslo_concurrency.lockutils [req-5103e08b-bfa9-4783-97ab-5a49959739a1 req-af2ee6dc-7ed1-4d48-a379-9c2e8e2bcc70 service nova] Lock "0416ef03-fd45-405f-9047-76b34a839217-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.190969] env[62600]: DEBUG nova.compute.manager [req-5103e08b-bfa9-4783-97ab-5a49959739a1 req-af2ee6dc-7ed1-4d48-a379-9c2e8e2bcc70 service nova] [instance: 0416ef03-fd45-405f-9047-76b34a839217] No waiting events found dispatching network-vif-plugged-bb15b28d-be8f-4fff-b8a6-0801d3e0541a {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1009.192805] env[62600]: WARNING nova.compute.manager [req-5103e08b-bfa9-4783-97ab-5a49959739a1 req-af2ee6dc-7ed1-4d48-a379-9c2e8e2bcc70 service nova] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Received unexpected event network-vif-plugged-bb15b28d-be8f-4fff-b8a6-0801d3e0541a for instance with vm_state building and task_state spawning. [ 1009.248207] env[62600]: DEBUG oslo_concurrency.lockutils [None req-65e09110-e1b6-48b3-9a9a-b8571fcbb519 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "39deb498-6bf0-4f3b-932b-8068fc48271e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.636s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.272630] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222931, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448943} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.273017] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] a20222a7-f29b-468f-8105-50ff9e31bab2/a20222a7-f29b-468f-8105-50ff9e31bab2.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1009.273350] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1009.273648] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1d9673d1-cd56-448b-8ae8-db3e913d3cee {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.282198] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1009.282198] env[62600]: value = "task-1222932" [ 1009.282198] env[62600]: _type = "Task" [ 1009.282198] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.282702] env[62600]: DEBUG nova.network.neutron [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Successfully updated port: bb15b28d-be8f-4fff-b8a6-0801d3e0541a {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1009.292945] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222932, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.431192] env[62600]: DEBUG oslo_vmware.api [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] Task: {'id': task-1222926, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.788999] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "refresh_cache-0416ef03-fd45-405f-9047-76b34a839217" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.789281] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "refresh_cache-0416ef03-fd45-405f-9047-76b34a839217" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.789509] env[62600]: DEBUG nova.network.neutron [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1009.796846] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222932, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065827} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.797154] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1009.797962] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3bdc62e-b62f-4a4d-bb9f-5a660ffe4e7a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.820922] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] a20222a7-f29b-468f-8105-50ff9e31bab2/a20222a7-f29b-468f-8105-50ff9e31bab2.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1009.821552] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12ce5bb7-c9b2-4e7e-8b37-e67d513b8d74 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.836287] env[62600]: DEBUG nova.network.neutron [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Updating instance_info_cache with network_info: [{"id": "7af8f843-5635-4c50-bc15-f8356b10e6f7", "address": "fa:16:3e:fc:4b:8b", "network": {"id": "5ef67e8a-5613-4cdb-9f35-0a18720657ff", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1644754500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bab74e3ae78248909dbcd483abbb8da2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7af8f843-56", "ovs_interfaceid": "7af8f843-5635-4c50-bc15-f8356b10e6f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.844130] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1009.844130] env[62600]: value = "task-1222933" [ 1009.844130] env[62600]: _type = "Task" [ 1009.844130] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.856703] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222933, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.931264] env[62600]: DEBUG oslo_vmware.api [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] Task: {'id': task-1222926, 'name': ReconfigVM_Task, 'duration_secs': 5.789704} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.931572] env[62600]: DEBUG oslo_concurrency.lockutils [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] Releasing lock "88cf521d-f4a3-493a-894d-c7f799dc0443" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.931726] env[62600]: DEBUG nova.virt.vmwareapi.vmops [req-b0b013b4-52e0-45af-a57d-e615b7f6df5c req-28460857-f3e3-485c-95f0-25ef0fadd004 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Reconfigured VM to detach interface {{(pid=62600) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1009.932224] env[62600]: DEBUG oslo_concurrency.lockutils [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "88cf521d-f4a3-493a-894d-c7f799dc0443" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 6.097s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.932504] env[62600]: DEBUG oslo_concurrency.lockutils [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "88cf521d-f4a3-493a-894d-c7f799dc0443-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.932721] env[62600]: DEBUG oslo_concurrency.lockutils [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "88cf521d-f4a3-493a-894d-c7f799dc0443-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.932901] env[62600]: DEBUG oslo_concurrency.lockutils [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "88cf521d-f4a3-493a-894d-c7f799dc0443-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.935022] env[62600]: INFO nova.compute.manager [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Terminating instance [ 1009.936913] env[62600]: DEBUG nova.compute.manager [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1009.937122] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1009.938271] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c71a8d-6a97-4db8-8879-f85703442219 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.946472] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1009.946743] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b412ee2a-b0a5-4835-be0d-358ac50d6f74 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.953821] env[62600]: DEBUG oslo_vmware.api [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1009.953821] env[62600]: value = "task-1222934" [ 1009.953821] env[62600]: _type = "Task" [ 1009.953821] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.963171] env[62600]: DEBUG oslo_vmware.api [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222934, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.326886] env[62600]: DEBUG nova.network.neutron [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1010.338756] env[62600]: DEBUG oslo_concurrency.lockutils [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Releasing lock "refresh_cache-aff91190-e19b-4f1c-8b47-9539360a4596" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.356686] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222933, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.467650] env[62600]: DEBUG oslo_vmware.api [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222934, 'name': PowerOffVM_Task, 'duration_secs': 0.199009} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.467650] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1010.467650] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1010.467650] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e40f4f0f-d963-425c-8e1a-db48d5f05c17 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.472426] env[62600]: DEBUG nova.network.neutron [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Updating instance_info_cache with network_info: [{"id": "bb15b28d-be8f-4fff-b8a6-0801d3e0541a", "address": "fa:16:3e:b1:43:47", "network": {"id": "95818650-8227-4d11-9d14-007f06c07790", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1738152115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73e664d178f7484a9f4741b4d9450e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb15b28d-be", "ovs_interfaceid": "bb15b28d-be8f-4fff-b8a6-0801d3e0541a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.565199] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1010.565546] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1010.565799] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Deleting the datastore file [datastore2] 88cf521d-f4a3-493a-894d-c7f799dc0443 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1010.566158] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ad4490b8-59ab-42a6-9d4b-71b3b949eaaa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.574199] env[62600]: DEBUG oslo_vmware.api [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1010.574199] env[62600]: value = "task-1222936" [ 1010.574199] env[62600]: _type = "Task" [ 1010.574199] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.582024] env[62600]: DEBUG oslo_vmware.api [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222936, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.836539] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f6ee0da0-90bf-48df-a296-520b150f3df9 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "ab11cdd1-80a7-43c4-afac-24c7930238c1" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.836878] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f6ee0da0-90bf-48df-a296-520b150f3df9 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "ab11cdd1-80a7-43c4-afac-24c7930238c1" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.861203] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222933, 'name': ReconfigVM_Task, 'duration_secs': 0.900139} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.861492] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Reconfigured VM instance instance-0000005f to attach disk [datastore2] a20222a7-f29b-468f-8105-50ff9e31bab2/a20222a7-f29b-468f-8105-50ff9e31bab2.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1010.862315] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eae675ec-46f9-4c0c-950e-1ce3d3da414f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.868988] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1feb4c8d-1f98-45ec-89a5-ccf643478692 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.873225] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1010.873225] env[62600]: value = "task-1222937" [ 1010.873225] env[62600]: _type = "Task" [ 1010.873225] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.894441] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2570b9ef-cde6-4d6e-8174-c106350e53e4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.900819] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222937, 'name': Rename_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.905748] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Updating instance 'aff91190-e19b-4f1c-8b47-9539360a4596' progress to 83 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1010.975060] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "refresh_cache-0416ef03-fd45-405f-9047-76b34a839217" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.975442] env[62600]: DEBUG nova.compute.manager [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Instance network_info: |[{"id": "bb15b28d-be8f-4fff-b8a6-0801d3e0541a", "address": "fa:16:3e:b1:43:47", "network": {"id": "95818650-8227-4d11-9d14-007f06c07790", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1738152115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73e664d178f7484a9f4741b4d9450e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb15b28d-be", "ovs_interfaceid": "bb15b28d-be8f-4fff-b8a6-0801d3e0541a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1010.975920] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b1:43:47', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43798f54-0c07-4417-a23f-58bb6b7e204b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bb15b28d-be8f-4fff-b8a6-0801d3e0541a', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1010.985167] env[62600]: DEBUG oslo.service.loopingcall [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1010.985428] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1010.985672] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8f6cfcbe-649d-423f-a49c-f64a862fd5f3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.006429] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1011.006429] env[62600]: value = "task-1222938" [ 1011.006429] env[62600]: _type = "Task" [ 1011.006429] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.014315] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222938, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.084464] env[62600]: DEBUG oslo_vmware.api [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222936, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.360761} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.084879] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1011.085138] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1011.085372] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1011.085570] env[62600]: INFO nova.compute.manager [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1011.085833] env[62600]: DEBUG oslo.service.loopingcall [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.086052] env[62600]: DEBUG nova.compute.manager [-] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1011.086153] env[62600]: DEBUG nova.network.neutron [-] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1011.163561] env[62600]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 0bc0b585-65b4-4ebf-81be-bbc628771202 could not be found.", "detail": ""}} {{(pid=62600) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 1011.163811] env[62600]: DEBUG nova.network.neutron [-] Unable to show port 0bc0b585-65b4-4ebf-81be-bbc628771202 as it no longer exists. {{(pid=62600) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 1011.219989] env[62600]: DEBUG nova.compute.manager [req-af26f8ec-28d8-4ba7-bbcf-ccba3b0f71ef req-ef57af56-a772-46e1-8180-0aa659544622 service nova] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Received event network-changed-bb15b28d-be8f-4fff-b8a6-0801d3e0541a {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.220313] env[62600]: DEBUG nova.compute.manager [req-af26f8ec-28d8-4ba7-bbcf-ccba3b0f71ef req-ef57af56-a772-46e1-8180-0aa659544622 service nova] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Refreshing instance network info cache due to event network-changed-bb15b28d-be8f-4fff-b8a6-0801d3e0541a. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1011.220557] env[62600]: DEBUG oslo_concurrency.lockutils [req-af26f8ec-28d8-4ba7-bbcf-ccba3b0f71ef req-ef57af56-a772-46e1-8180-0aa659544622 service nova] Acquiring lock "refresh_cache-0416ef03-fd45-405f-9047-76b34a839217" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.220717] env[62600]: DEBUG oslo_concurrency.lockutils [req-af26f8ec-28d8-4ba7-bbcf-ccba3b0f71ef req-ef57af56-a772-46e1-8180-0aa659544622 service nova] Acquired lock "refresh_cache-0416ef03-fd45-405f-9047-76b34a839217" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.220905] env[62600]: DEBUG nova.network.neutron [req-af26f8ec-28d8-4ba7-bbcf-ccba3b0f71ef req-ef57af56-a772-46e1-8180-0aa659544622 service nova] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Refreshing network info cache for port bb15b28d-be8f-4fff-b8a6-0801d3e0541a {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1011.339762] env[62600]: INFO nova.compute.manager [None req-f6ee0da0-90bf-48df-a296-520b150f3df9 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Detaching volume c49c9652-3f74-46a1-a176-6b393fc416ba [ 1011.378658] env[62600]: INFO nova.virt.block_device [None req-f6ee0da0-90bf-48df-a296-520b150f3df9 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Attempting to driver detach volume c49c9652-3f74-46a1-a176-6b393fc416ba from mountpoint /dev/sdb [ 1011.378932] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6ee0da0-90bf-48df-a296-520b150f3df9 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Volume detach. Driver type: vmdk {{(pid=62600) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1011.379151] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6ee0da0-90bf-48df-a296-520b150f3df9 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264337', 'volume_id': 'c49c9652-3f74-46a1-a176-6b393fc416ba', 'name': 'volume-c49c9652-3f74-46a1-a176-6b393fc416ba', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ab11cdd1-80a7-43c4-afac-24c7930238c1', 'attached_at': '', 'detached_at': '', 'volume_id': 'c49c9652-3f74-46a1-a176-6b393fc416ba', 'serial': 'c49c9652-3f74-46a1-a176-6b393fc416ba'} {{(pid=62600) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1011.379990] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d8e1fe0-d362-421f-8e6d-942461111402 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.389043] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222937, 'name': Rename_Task, 'duration_secs': 0.143255} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.404517] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1011.405017] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1181c69f-ffae-4448-a672-785c493d7e14 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.407425] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61ac61eb-5448-4a89-ad30-b67067bd172f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.411636] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1011.412327] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6d4f9eb8-464e-4ab0-baec-05db935934f3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.418356] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1011.418356] env[62600]: value = "task-1222939" [ 1011.418356] env[62600]: _type = "Task" [ 1011.418356] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.428407] env[62600]: DEBUG oslo_vmware.api [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 1011.428407] env[62600]: value = "task-1222940" [ 1011.428407] env[62600]: _type = "Task" [ 1011.428407] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.430384] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce54e44-abf0-4735-bb5a-7c556e698597 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.442081] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222939, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.468580] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735dcc20-bd8f-4174-8136-4e73373889be {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.472099] env[62600]: DEBUG oslo_vmware.api [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222940, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.487100] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6ee0da0-90bf-48df-a296-520b150f3df9 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] The volume has not been displaced from its original location: [datastore2] volume-c49c9652-3f74-46a1-a176-6b393fc416ba/volume-c49c9652-3f74-46a1-a176-6b393fc416ba.vmdk. No consolidation needed. {{(pid=62600) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1011.493154] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6ee0da0-90bf-48df-a296-520b150f3df9 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Reconfiguring VM instance instance-00000046 to detach disk 2001 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1011.493883] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7f70bed-de4a-47ef-adb5-59f1c6a91adc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.517815] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222938, 'name': CreateVM_Task, 'duration_secs': 0.476434} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.519083] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1011.519496] env[62600]: DEBUG oslo_vmware.api [None req-f6ee0da0-90bf-48df-a296-520b150f3df9 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1011.519496] env[62600]: value = "task-1222941" [ 1011.519496] env[62600]: _type = "Task" [ 1011.519496] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.520262] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.520476] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.520885] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1011.521276] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19dceedb-4ed1-4324-8a98-8c1597ed8ea2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.531582] env[62600]: DEBUG oslo_vmware.api [None req-f6ee0da0-90bf-48df-a296-520b150f3df9 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222941, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.532636] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1011.532636] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c9f5ab-8a57-ea03-4c64-ed24826493d2" [ 1011.532636] env[62600]: _type = "Task" [ 1011.532636] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.541508] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c9f5ab-8a57-ea03-4c64-ed24826493d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.928587] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222939, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.935471] env[62600]: DEBUG nova.network.neutron [req-af26f8ec-28d8-4ba7-bbcf-ccba3b0f71ef req-ef57af56-a772-46e1-8180-0aa659544622 service nova] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Updated VIF entry in instance network info cache for port bb15b28d-be8f-4fff-b8a6-0801d3e0541a. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1011.935948] env[62600]: DEBUG nova.network.neutron [req-af26f8ec-28d8-4ba7-bbcf-ccba3b0f71ef req-ef57af56-a772-46e1-8180-0aa659544622 service nova] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Updating instance_info_cache with network_info: [{"id": "bb15b28d-be8f-4fff-b8a6-0801d3e0541a", "address": "fa:16:3e:b1:43:47", "network": {"id": "95818650-8227-4d11-9d14-007f06c07790", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1738152115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73e664d178f7484a9f4741b4d9450e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb15b28d-be", "ovs_interfaceid": "bb15b28d-be8f-4fff-b8a6-0801d3e0541a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.947593] env[62600]: DEBUG oslo_vmware.api [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222940, 'name': PowerOnVM_Task, 'duration_secs': 0.464477} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.947886] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1011.948106] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-ca497c93-68ec-43ed-a664-b50d7d72aed5 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Updating instance 'aff91190-e19b-4f1c-8b47-9539360a4596' progress to 100 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1012.031210] env[62600]: DEBUG oslo_vmware.api [None req-f6ee0da0-90bf-48df-a296-520b150f3df9 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222941, 'name': ReconfigVM_Task, 'duration_secs': 0.300268} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.031474] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6ee0da0-90bf-48df-a296-520b150f3df9 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Reconfigured VM instance instance-00000046 to detach disk 2001 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1012.036113] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec908851-eea9-4ece-8af7-b78955009c87 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.054842] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c9f5ab-8a57-ea03-4c64-ed24826493d2, 'name': SearchDatastore_Task, 'duration_secs': 0.01157} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.056199] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.056448] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1012.056686] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.056839] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.057032] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1012.057355] env[62600]: DEBUG oslo_vmware.api [None req-f6ee0da0-90bf-48df-a296-520b150f3df9 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1012.057355] env[62600]: value = "task-1222942" [ 1012.057355] env[62600]: _type = "Task" [ 1012.057355] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.057561] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-22b948b0-15f5-450f-b00b-d0b30deda5d9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.062689] env[62600]: DEBUG nova.network.neutron [-] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.069705] env[62600]: DEBUG oslo_vmware.api [None req-f6ee0da0-90bf-48df-a296-520b150f3df9 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222942, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.070880] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1012.071077] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1012.071812] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57e1b371-ac19-4ba3-9055-bfd90c38fdb5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.077867] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1012.077867] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5244e7a4-eb9a-5965-c60a-573a61966096" [ 1012.077867] env[62600]: _type = "Task" [ 1012.077867] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.086797] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5244e7a4-eb9a-5965-c60a-573a61966096, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.431290] env[62600]: DEBUG oslo_vmware.api [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222939, 'name': PowerOnVM_Task, 'duration_secs': 0.75403} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.431581] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1012.431773] env[62600]: INFO nova.compute.manager [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Took 7.96 seconds to spawn the instance on the hypervisor. [ 1012.431959] env[62600]: DEBUG nova.compute.manager [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1012.432731] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc35af56-c5f7-45ac-849d-4e2d3aadf08b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.441523] env[62600]: DEBUG oslo_concurrency.lockutils [req-af26f8ec-28d8-4ba7-bbcf-ccba3b0f71ef req-ef57af56-a772-46e1-8180-0aa659544622 service nova] Releasing lock "refresh_cache-0416ef03-fd45-405f-9047-76b34a839217" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.564930] env[62600]: INFO nova.compute.manager [-] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Took 1.48 seconds to deallocate network for instance. [ 1012.569667] env[62600]: DEBUG oslo_vmware.api [None req-f6ee0da0-90bf-48df-a296-520b150f3df9 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222942, 'name': ReconfigVM_Task, 'duration_secs': 0.211313} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.572460] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6ee0da0-90bf-48df-a296-520b150f3df9 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264337', 'volume_id': 'c49c9652-3f74-46a1-a176-6b393fc416ba', 'name': 'volume-c49c9652-3f74-46a1-a176-6b393fc416ba', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ab11cdd1-80a7-43c4-afac-24c7930238c1', 'attached_at': '', 'detached_at': '', 'volume_id': 'c49c9652-3f74-46a1-a176-6b393fc416ba', 'serial': 'c49c9652-3f74-46a1-a176-6b393fc416ba'} {{(pid=62600) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1012.588597] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5244e7a4-eb9a-5965-c60a-573a61966096, 'name': SearchDatastore_Task, 'duration_secs': 0.013114} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.589375] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7eb92d0b-a4b5-4167-9641-918e4f682828 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.594772] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1012.594772] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5208e11b-d99a-9fb5-1744-946ea403bc8a" [ 1012.594772] env[62600]: _type = "Task" [ 1012.594772] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.603091] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5208e11b-d99a-9fb5-1744-946ea403bc8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.950705] env[62600]: INFO nova.compute.manager [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Took 14.40 seconds to build instance. [ 1013.076391] env[62600]: DEBUG oslo_concurrency.lockutils [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.076669] env[62600]: DEBUG oslo_concurrency.lockutils [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.076900] env[62600]: DEBUG nova.objects.instance [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lazy-loading 'resources' on Instance uuid 88cf521d-f4a3-493a-894d-c7f799dc0443 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.105390] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5208e11b-d99a-9fb5-1744-946ea403bc8a, 'name': SearchDatastore_Task, 'duration_secs': 0.009601} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.105720] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.105902] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 0416ef03-fd45-405f-9047-76b34a839217/0416ef03-fd45-405f-9047-76b34a839217.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1013.106180] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a96ad129-736c-4911-88d5-987d2c6e4ba5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.114244] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1013.114244] env[62600]: value = "task-1222943" [ 1013.114244] env[62600]: _type = "Task" [ 1013.114244] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.115561] env[62600]: DEBUG nova.objects.instance [None req-f6ee0da0-90bf-48df-a296-520b150f3df9 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lazy-loading 'flavor' on Instance uuid ab11cdd1-80a7-43c4-afac-24c7930238c1 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.126595] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222943, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.247077] env[62600]: DEBUG nova.compute.manager [req-60f0cc30-7d3d-44e1-af41-54af544a41fb req-87af761b-16fd-4985-86e3-c142556e0356 service nova] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Received event network-vif-deleted-3a8aca44-9791-4493-a1fd-114c9f20fb3f {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1013.453637] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8a309f23-4245-4898-a036-a1c2d4401964 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "a20222a7-f29b-468f-8105-50ff9e31bab2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.910s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.528868] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "a20222a7-f29b-468f-8105-50ff9e31bab2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.529316] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "a20222a7-f29b-468f-8105-50ff9e31bab2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.529588] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "a20222a7-f29b-468f-8105-50ff9e31bab2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.529904] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "a20222a7-f29b-468f-8105-50ff9e31bab2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.530217] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "a20222a7-f29b-468f-8105-50ff9e31bab2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.533173] env[62600]: INFO nova.compute.manager [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Terminating instance [ 1013.535453] env[62600]: DEBUG nova.compute.manager [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1013.535737] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1013.536874] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e43617c-36ee-4a29-8bf7-b45cce9a716b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.547075] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1013.547426] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e8256d6f-7e4c-4831-981d-bfc97db14006 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.557052] env[62600]: DEBUG oslo_vmware.api [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1013.557052] env[62600]: value = "task-1222944" [ 1013.557052] env[62600]: _type = "Task" [ 1013.557052] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.569113] env[62600]: DEBUG oslo_vmware.api [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222944, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.631083] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222943, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478918} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.633721] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 0416ef03-fd45-405f-9047-76b34a839217/0416ef03-fd45-405f-9047-76b34a839217.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1013.633952] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1013.634889] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a43a895-7000-4269-acb2-42f9eca96daa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.642930] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1013.642930] env[62600]: value = "task-1222945" [ 1013.642930] env[62600]: _type = "Task" [ 1013.642930] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.654540] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222945, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.762678] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f6ff698-6f0c-400a-94d6-37f223366ee7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.770594] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e7a8e65-4248-48a1-ad7d-b8f2a42761b1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.800190] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a94c6f-99b6-4a98-8dab-d4561eda2d66 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.807531] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d72ed2-bd97-4975-8ac3-0b210806c89d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.820517] env[62600]: DEBUG nova.compute.provider_tree [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1014.068992] env[62600]: DEBUG oslo_vmware.api [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222944, 'name': PowerOffVM_Task, 'duration_secs': 0.204825} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.069396] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1014.069470] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1014.069730] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c15c92c-f6fd-402e-8bba-5f56bf79a9eb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.086291] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "aff91190-e19b-4f1c-8b47-9539360a4596" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.086556] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "aff91190-e19b-4f1c-8b47-9539360a4596" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.086760] env[62600]: DEBUG nova.compute.manager [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Going to confirm migration 3 {{(pid=62600) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1014.126295] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f6ee0da0-90bf-48df-a296-520b150f3df9 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "ab11cdd1-80a7-43c4-afac-24c7930238c1" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.289s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.152487] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222945, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067166} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.152767] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1014.153559] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2405ac92-59da-470a-bd21-4c6d04b1c6a3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.175414] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 0416ef03-fd45-405f-9047-76b34a839217/0416ef03-fd45-405f-9047-76b34a839217.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1014.175993] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d6e358f7-8134-4a4a-8764-797630c8adea {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.196425] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1014.196425] env[62600]: value = "task-1222947" [ 1014.196425] env[62600]: _type = "Task" [ 1014.196425] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.206039] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222947, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.350459] env[62600]: DEBUG nova.scheduler.client.report [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 130 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1014.350793] env[62600]: DEBUG nova.compute.provider_tree [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 130 to 131 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1014.351057] env[62600]: DEBUG nova.compute.provider_tree [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1014.363110] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1014.363346] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1014.363540] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleting the datastore file [datastore2] a20222a7-f29b-468f-8105-50ff9e31bab2 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1014.363805] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ed92d3d8-1edc-423f-ace9-62ba0c62f8fa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.375300] env[62600]: DEBUG oslo_vmware.api [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1014.375300] env[62600]: value = "task-1222948" [ 1014.375300] env[62600]: _type = "Task" [ 1014.375300] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.384423] env[62600]: DEBUG oslo_vmware.api [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222948, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.651343] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "refresh_cache-aff91190-e19b-4f1c-8b47-9539360a4596" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.651584] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquired lock "refresh_cache-aff91190-e19b-4f1c-8b47-9539360a4596" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.651839] env[62600]: DEBUG nova.network.neutron [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1014.652118] env[62600]: DEBUG nova.objects.instance [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lazy-loading 'info_cache' on Instance uuid aff91190-e19b-4f1c-8b47-9539360a4596 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1014.706979] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222947, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.857065] env[62600]: DEBUG oslo_concurrency.lockutils [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.780s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.880750] env[62600]: INFO nova.scheduler.client.report [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Deleted allocations for instance 88cf521d-f4a3-493a-894d-c7f799dc0443 [ 1014.887712] env[62600]: DEBUG oslo_vmware.api [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222948, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137506} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.888189] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1014.888390] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1014.888575] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1014.888753] env[62600]: INFO nova.compute.manager [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Took 1.35 seconds to destroy the instance on the hypervisor. [ 1014.889007] env[62600]: DEBUG oslo.service.loopingcall [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.889361] env[62600]: DEBUG nova.compute.manager [-] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1014.889469] env[62600]: DEBUG nova.network.neutron [-] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1015.150268] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "ab11cdd1-80a7-43c4-afac-24c7930238c1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.150424] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "ab11cdd1-80a7-43c4-afac-24c7930238c1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.150645] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "ab11cdd1-80a7-43c4-afac-24c7930238c1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.150839] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "ab11cdd1-80a7-43c4-afac-24c7930238c1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.151027] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "ab11cdd1-80a7-43c4-afac-24c7930238c1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.153328] env[62600]: INFO nova.compute.manager [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Terminating instance [ 1015.156669] env[62600]: DEBUG nova.compute.manager [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1015.156871] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1015.157915] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-966264e9-f1e8-402c-ab3d-94329920577b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.171492] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1015.171656] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3670b5e7-db3f-447c-87cf-8297737e2a56 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.179293] env[62600]: DEBUG oslo_vmware.api [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1015.179293] env[62600]: value = "task-1222949" [ 1015.179293] env[62600]: _type = "Task" [ 1015.179293] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.188630] env[62600]: DEBUG oslo_vmware.api [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222949, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.209037] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222947, 'name': ReconfigVM_Task, 'duration_secs': 0.933933} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.209352] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 0416ef03-fd45-405f-9047-76b34a839217/0416ef03-fd45-405f-9047-76b34a839217.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1015.210439] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0579199c-9607-481c-9374-6f7a8e9ddb7b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.216806] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1015.216806] env[62600]: value = "task-1222950" [ 1015.216806] env[62600]: _type = "Task" [ 1015.216806] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.225198] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222950, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.377375] env[62600]: DEBUG nova.network.neutron [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Updating instance_info_cache with network_info: [{"id": "7af8f843-5635-4c50-bc15-f8356b10e6f7", "address": "fa:16:3e:fc:4b:8b", "network": {"id": "5ef67e8a-5613-4cdb-9f35-0a18720657ff", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1644754500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bab74e3ae78248909dbcd483abbb8da2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7af8f843-56", "ovs_interfaceid": "7af8f843-5635-4c50-bc15-f8356b10e6f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.381801] env[62600]: DEBUG nova.compute.manager [req-f85fe0a6-e25c-485c-9a78-9eeef1281857 req-22317981-d597-4b8d-9d53-f1ca74c68d58 service nova] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Received event network-vif-deleted-0ffa19db-f509-4c53-be0e-7a1692d2f4b7 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1015.381801] env[62600]: INFO nova.compute.manager [req-f85fe0a6-e25c-485c-9a78-9eeef1281857 req-22317981-d597-4b8d-9d53-f1ca74c68d58 service nova] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Neutron deleted interface 0ffa19db-f509-4c53-be0e-7a1692d2f4b7; detaching it from the instance and deleting it from the info cache [ 1015.381801] env[62600]: DEBUG nova.network.neutron [req-f85fe0a6-e25c-485c-9a78-9eeef1281857 req-22317981-d597-4b8d-9d53-f1ca74c68d58 service nova] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.391309] env[62600]: DEBUG oslo_concurrency.lockutils [None req-adef1e9f-d8f2-4a09-9355-a295e0bdeb4b tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "88cf521d-f4a3-493a-894d-c7f799dc0443" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.459s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.643035] env[62600]: DEBUG nova.network.neutron [-] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.689864] env[62600]: DEBUG oslo_vmware.api [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222949, 'name': PowerOffVM_Task, 'duration_secs': 0.423284} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.690185] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1015.690344] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1015.690611] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-872b71fb-d904-443a-9d72-36e3d3c25ce7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.727750] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222950, 'name': Rename_Task, 'duration_secs': 0.359668} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.728497] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1015.728782] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-34dd36d7-f362-4784-8dd3-c69bc55b1fa5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.736060] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1015.736060] env[62600]: value = "task-1222952" [ 1015.736060] env[62600]: _type = "Task" [ 1015.736060] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.744750] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222952, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.767644] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1015.767896] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1015.768101] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Deleting the datastore file [datastore1] ab11cdd1-80a7-43c4-afac-24c7930238c1 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1015.768409] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-41fe4b2d-2898-4dbb-a3bb-ad8656f18290 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.776810] env[62600]: DEBUG oslo_vmware.api [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1015.776810] env[62600]: value = "task-1222953" [ 1015.776810] env[62600]: _type = "Task" [ 1015.776810] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.786437] env[62600]: DEBUG oslo_vmware.api [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222953, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.817169] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4e13b314-4fea-4c1b-820d-8a1f39321073 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "248a5371-6ff7-4da7-ae97-d638bea0123a" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.817528] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4e13b314-4fea-4c1b-820d-8a1f39321073 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "248a5371-6ff7-4da7-ae97-d638bea0123a" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.881632] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Releasing lock "refresh_cache-aff91190-e19b-4f1c-8b47-9539360a4596" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.881881] env[62600]: DEBUG nova.objects.instance [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lazy-loading 'migration_context' on Instance uuid aff91190-e19b-4f1c-8b47-9539360a4596 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.884118] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ff634aab-75e9-4dd3-99df-93091ae60b1f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.895771] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c3150c-89dc-4d50-91e6-90db4d6b4512 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.928181] env[62600]: DEBUG nova.compute.manager [req-f85fe0a6-e25c-485c-9a78-9eeef1281857 req-22317981-d597-4b8d-9d53-f1ca74c68d58 service nova] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Detach interface failed, port_id=0ffa19db-f509-4c53-be0e-7a1692d2f4b7, reason: Instance a20222a7-f29b-468f-8105-50ff9e31bab2 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1016.145376] env[62600]: INFO nova.compute.manager [-] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Took 1.26 seconds to deallocate network for instance. [ 1016.247526] env[62600]: DEBUG oslo_vmware.api [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222952, 'name': PowerOnVM_Task, 'duration_secs': 0.486838} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.247888] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1016.248169] env[62600]: INFO nova.compute.manager [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Took 7.27 seconds to spawn the instance on the hypervisor. [ 1016.248397] env[62600]: DEBUG nova.compute.manager [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1016.249224] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d1f438e-7715-4cf2-aeba-78e216e1d713 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.287604] env[62600]: DEBUG oslo_vmware.api [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222953, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162797} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.287895] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1016.288099] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1016.288294] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1016.288482] env[62600]: INFO nova.compute.manager [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1016.288733] env[62600]: DEBUG oslo.service.loopingcall [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1016.288931] env[62600]: DEBUG nova.compute.manager [-] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1016.289039] env[62600]: DEBUG nova.network.neutron [-] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1016.320670] env[62600]: INFO nova.compute.manager [None req-4e13b314-4fea-4c1b-820d-8a1f39321073 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Detaching volume e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf [ 1016.357047] env[62600]: INFO nova.virt.block_device [None req-4e13b314-4fea-4c1b-820d-8a1f39321073 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Attempting to driver detach volume e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf from mountpoint /dev/sdb [ 1016.357309] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e13b314-4fea-4c1b-820d-8a1f39321073 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Volume detach. Driver type: vmdk {{(pid=62600) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1016.357513] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e13b314-4fea-4c1b-820d-8a1f39321073 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264349', 'volume_id': 'e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf', 'name': 'volume-e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '248a5371-6ff7-4da7-ae97-d638bea0123a', 'attached_at': '', 'detached_at': '', 'volume_id': 'e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf', 'serial': 'e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf'} {{(pid=62600) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1016.358396] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d264f63-4219-44f7-8e7c-c9bb502a4df6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.380469] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-916d61b4-f347-4d8c-9038-3b682f246a3b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.385186] env[62600]: DEBUG nova.objects.base [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62600) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1016.386053] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca160dbb-e5c4-49ef-baa4-24a4aca634dc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.390934] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe65fa4-f42f-4a1b-83a5-c536ca267063 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.407554] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f6bb941-ef36-45c1-bb2b-333ca37a28de {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.430574] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f019911e-5287-44ab-a2ff-a75cd443c141 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.433354] env[62600]: DEBUG oslo_vmware.api [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 1016.433354] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528df85a-a08e-0149-c7f4-40d91b3f64d2" [ 1016.433354] env[62600]: _type = "Task" [ 1016.433354] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.447427] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e13b314-4fea-4c1b-820d-8a1f39321073 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] The volume has not been displaced from its original location: [datastore1] volume-e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf/volume-e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf.vmdk. No consolidation needed. {{(pid=62600) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1016.452685] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e13b314-4fea-4c1b-820d-8a1f39321073 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Reconfiguring VM instance instance-00000050 to detach disk 2001 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1016.453503] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db061978-714f-41c6-b633-1c71dba88c7d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.470277] env[62600]: DEBUG oslo_vmware.api [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528df85a-a08e-0149-c7f4-40d91b3f64d2, 'name': SearchDatastore_Task, 'duration_secs': 0.007561} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.470965] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.471223] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.476718] env[62600]: DEBUG oslo_vmware.api [None req-4e13b314-4fea-4c1b-820d-8a1f39321073 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 1016.476718] env[62600]: value = "task-1222954" [ 1016.476718] env[62600]: _type = "Task" [ 1016.476718] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.485973] env[62600]: DEBUG oslo_vmware.api [None req-4e13b314-4fea-4c1b-820d-8a1f39321073 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222954, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.654608] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.772494] env[62600]: INFO nova.compute.manager [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Took 12.98 seconds to build instance. [ 1016.986902] env[62600]: DEBUG oslo_vmware.api [None req-4e13b314-4fea-4c1b-820d-8a1f39321073 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222954, 'name': ReconfigVM_Task, 'duration_secs': 0.243545} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.987273] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e13b314-4fea-4c1b-820d-8a1f39321073 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Reconfigured VM instance instance-00000050 to detach disk 2001 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1016.992667] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ad2d5eb-8227-4520-bfe4-f8d8310bbd9f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.011359] env[62600]: DEBUG oslo_vmware.api [None req-4e13b314-4fea-4c1b-820d-8a1f39321073 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 1017.011359] env[62600]: value = "task-1222955" [ 1017.011359] env[62600]: _type = "Task" [ 1017.011359] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.020687] env[62600]: DEBUG oslo_vmware.api [None req-4e13b314-4fea-4c1b-820d-8a1f39321073 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222955, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.147391] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-221cd908-2c43-426b-bd35-b5f0541dadaa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.155401] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4bf3e4f-3c1e-47c8-80c0-6438170f7209 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.190370] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58cf2e0d-1606-420f-807e-a5a5c842973c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.198771] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-743ac24c-a478-4e35-9458-a8ff3a6eb831 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.213291] env[62600]: DEBUG nova.network.neutron [-] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.214982] env[62600]: DEBUG nova.compute.provider_tree [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1017.275409] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9590aeb7-bb09-4577-b5b3-ea3dc26e1f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "0416ef03-fd45-405f-9047-76b34a839217" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.487s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.522548] env[62600]: DEBUG oslo_vmware.api [None req-4e13b314-4fea-4c1b-820d-8a1f39321073 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222955, 'name': ReconfigVM_Task, 'duration_secs': 0.152806} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.523207] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e13b314-4fea-4c1b-820d-8a1f39321073 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264349', 'volume_id': 'e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf', 'name': 'volume-e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '248a5371-6ff7-4da7-ae97-d638bea0123a', 'attached_at': '', 'detached_at': '', 'volume_id': 'e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf', 'serial': 'e6f001d5-8622-4c3a-bfd6-e4068f5f8bcf'} {{(pid=62600) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1017.558929] env[62600]: DEBUG nova.compute.manager [req-2d5bfe1a-8931-4c2c-af5e-4f28cd117fd6 req-a33ecb30-aaea-4703-ace5-a195c7c7e020 service nova] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Received event network-vif-deleted-517fe831-7f4c-477a-9c43-496ec9895bb3 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1017.717522] env[62600]: INFO nova.compute.manager [-] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Took 1.43 seconds to deallocate network for instance. [ 1017.735636] env[62600]: ERROR nova.scheduler.client.report [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [req-693b5a66-62f3-448f-baad-e3471b9cd45e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 664af347-7147-4bf5-9019-9ae15cb4aa82. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-693b5a66-62f3-448f-baad-e3471b9cd45e"}]} [ 1017.752529] env[62600]: DEBUG nova.scheduler.client.report [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Refreshing inventories for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1017.767854] env[62600]: DEBUG nova.scheduler.client.report [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Updating ProviderTree inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1017.768153] env[62600]: DEBUG nova.compute.provider_tree [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1017.778621] env[62600]: DEBUG nova.scheduler.client.report [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Refreshing aggregate associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, aggregates: None {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1017.794848] env[62600]: DEBUG nova.scheduler.client.report [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Refreshing trait associations for resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE {{(pid=62600) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1017.928960] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52ebf933-85cc-4109-a13f-161ee4135ae8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.936364] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58212bb0-60b8-4b4f-870a-310f360d86cf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.967865] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee23c134-c38c-41af-ac6b-7e77529af29f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.975440] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-793af7b8-4fff-4c44-8eaa-0014d211dd51 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.989524] env[62600]: DEBUG nova.compute.provider_tree [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1018.043015] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "eb3c5624-0599-4e81-b056-bedf34329646" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.043368] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "eb3c5624-0599-4e81-b056-bedf34329646" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.070308] env[62600]: DEBUG nova.objects.instance [None req-4e13b314-4fea-4c1b-820d-8a1f39321073 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lazy-loading 'flavor' on Instance uuid 248a5371-6ff7-4da7-ae97-d638bea0123a {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.226881] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.298980] env[62600]: DEBUG nova.compute.manager [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Stashing vm_state: active {{(pid=62600) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1018.525653] env[62600]: DEBUG nova.scheduler.client.report [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Updated inventory for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with generation 134 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1018.525944] env[62600]: DEBUG nova.compute.provider_tree [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Updating resource provider 664af347-7147-4bf5-9019-9ae15cb4aa82 generation from 134 to 135 during operation: update_inventory {{(pid=62600) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1018.526159] env[62600]: DEBUG nova.compute.provider_tree [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Updating inventory in ProviderTree for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1018.545382] env[62600]: DEBUG nova.compute.manager [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1018.820277] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.063688] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.077634] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4e13b314-4fea-4c1b-820d-8a1f39321073 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "248a5371-6ff7-4da7-ae97-d638bea0123a" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.260s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.452985] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "248a5371-6ff7-4da7-ae97-d638bea0123a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.453325] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "248a5371-6ff7-4da7-ae97-d638bea0123a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.453549] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "248a5371-6ff7-4da7-ae97-d638bea0123a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.453740] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "248a5371-6ff7-4da7-ae97-d638bea0123a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.453916] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "248a5371-6ff7-4da7-ae97-d638bea0123a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.456342] env[62600]: INFO nova.compute.manager [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Terminating instance [ 1019.458158] env[62600]: DEBUG nova.compute.manager [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1019.458365] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1019.459217] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b455c2a-9d9e-4708-89da-1d63b417880b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.466931] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1019.467179] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-13a02be0-cb25-48cb-9529-90fa2eed9b79 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.475047] env[62600]: DEBUG oslo_vmware.api [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 1019.475047] env[62600]: value = "task-1222956" [ 1019.475047] env[62600]: _type = "Task" [ 1019.475047] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.482669] env[62600]: DEBUG oslo_vmware.api [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222956, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.536965] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 3.066s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.540233] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.886s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.540608] env[62600]: DEBUG nova.objects.instance [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lazy-loading 'resources' on Instance uuid a20222a7-f29b-468f-8105-50ff9e31bab2 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.985618] env[62600]: DEBUG oslo_vmware.api [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222956, 'name': PowerOffVM_Task, 'duration_secs': 0.230162} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.985951] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1019.986101] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1019.986336] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-74c07a7e-3274-47f8-b77f-213e298f9550 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.051681] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1020.051917] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1020.052160] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Deleting the datastore file [datastore1] 248a5371-6ff7-4da7-ae97-d638bea0123a {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1020.052604] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2dc9f6e8-b718-4c93-acf9-a54efc79c275 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.059426] env[62600]: DEBUG oslo_vmware.api [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 1020.059426] env[62600]: value = "task-1222958" [ 1020.059426] env[62600]: _type = "Task" [ 1020.059426] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.071192] env[62600]: DEBUG oslo_vmware.api [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222958, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.101538] env[62600]: INFO nova.scheduler.client.report [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Deleted allocation for migration e8e97f1c-3473-4838-bfb2-5e02bd200b70 [ 1020.204094] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7618df9-ae19-42a4-88e1-ce596e9bb91d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.211949] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19df7a5a-3a7f-46d9-ab1c-f052014f5f51 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.241020] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10aa5f8e-0c52-4e4c-82b1-6865b737a506 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.247951] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10209048-c70b-40c7-9f77-34f782463537 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.262073] env[62600]: DEBUG nova.compute.provider_tree [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1020.569313] env[62600]: DEBUG oslo_vmware.api [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222958, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155082} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.569573] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1020.569761] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1020.569944] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1020.570140] env[62600]: INFO nova.compute.manager [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1020.570390] env[62600]: DEBUG oslo.service.loopingcall [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.570585] env[62600]: DEBUG nova.compute.manager [-] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1020.570678] env[62600]: DEBUG nova.network.neutron [-] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1020.608250] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "aff91190-e19b-4f1c-8b47-9539360a4596" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.521s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.766525] env[62600]: DEBUG nova.scheduler.client.report [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1021.102681] env[62600]: DEBUG nova.compute.manager [req-12de5a2b-cef8-411d-ab9c-f63e7b086246 req-a23a603e-2b5e-414a-bb63-89d65a4c8b06 service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Received event network-vif-deleted-32f5d56d-a029-4469-ad3c-77a4ca51bc02 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1021.102956] env[62600]: INFO nova.compute.manager [req-12de5a2b-cef8-411d-ab9c-f63e7b086246 req-a23a603e-2b5e-414a-bb63-89d65a4c8b06 service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Neutron deleted interface 32f5d56d-a029-4469-ad3c-77a4ca51bc02; detaching it from the instance and deleting it from the info cache [ 1021.103229] env[62600]: DEBUG nova.network.neutron [req-12de5a2b-cef8-411d-ab9c-f63e7b086246 req-a23a603e-2b5e-414a-bb63-89d65a4c8b06 service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.270494] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.730s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.272943] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.046s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.273218] env[62600]: DEBUG nova.objects.instance [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lazy-loading 'resources' on Instance uuid ab11cdd1-80a7-43c4-afac-24c7930238c1 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1021.292147] env[62600]: INFO nova.scheduler.client.report [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleted allocations for instance a20222a7-f29b-468f-8105-50ff9e31bab2 [ 1021.515662] env[62600]: DEBUG nova.network.neutron [-] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.605890] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5f861611-70af-422b-ba71-c1c62b5561ac {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.616927] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae297fb-0919-4ed2-9594-ab17d445768e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.646791] env[62600]: DEBUG nova.compute.manager [req-12de5a2b-cef8-411d-ab9c-f63e7b086246 req-a23a603e-2b5e-414a-bb63-89d65a4c8b06 service nova] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Detach interface failed, port_id=32f5d56d-a029-4469-ad3c-77a4ca51bc02, reason: Instance 248a5371-6ff7-4da7-ae97-d638bea0123a could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1021.798682] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d1e3a389-ee4c-418d-800d-14916259342e tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "a20222a7-f29b-468f-8105-50ff9e31bab2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.269s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.905533] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-480fb1dd-c1d0-42c8-8032-9afbcc452024 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.913061] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a48cc5-074e-4301-af2e-0a0c912b69e9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.942242] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5780bd2-2177-4fae-a2ae-3e924b2ff0f9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.950026] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee8e1cc4-7eb8-4a33-94c3-0586473bd396 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.963192] env[62600]: DEBUG nova.compute.provider_tree [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.018637] env[62600]: INFO nova.compute.manager [-] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Took 1.45 seconds to deallocate network for instance. [ 1022.106783] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "aff91190-e19b-4f1c-8b47-9539360a4596" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.107116] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "aff91190-e19b-4f1c-8b47-9539360a4596" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.107363] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "aff91190-e19b-4f1c-8b47-9539360a4596-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.107592] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "aff91190-e19b-4f1c-8b47-9539360a4596-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.107790] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "aff91190-e19b-4f1c-8b47-9539360a4596-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.109996] env[62600]: INFO nova.compute.manager [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Terminating instance [ 1022.111880] env[62600]: DEBUG nova.compute.manager [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1022.112126] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1022.112997] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e301a30-7b97-46d6-9458-11719c0d82dc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.121517] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1022.122293] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3a9ab9f4-4e10-40b0-9d2b-1400372ad289 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.130979] env[62600]: DEBUG oslo_vmware.api [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 1022.130979] env[62600]: value = "task-1222959" [ 1022.130979] env[62600]: _type = "Task" [ 1022.130979] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.139101] env[62600]: DEBUG oslo_vmware.api [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222959, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.466145] env[62600]: DEBUG nova.scheduler.client.report [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1022.524960] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.640723] env[62600]: DEBUG oslo_vmware.api [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222959, 'name': PowerOffVM_Task, 'duration_secs': 0.276527} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.640976] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1022.641173] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1022.641436] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-30bc5a3e-7e21-4f53-abc2-8034d0b80100 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.685783] env[62600]: DEBUG oslo_concurrency.lockutils [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "5d7c2a17-9dee-407e-a7ee-6bb88f56905f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.686061] env[62600]: DEBUG oslo_concurrency.lockutils [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "5d7c2a17-9dee-407e-a7ee-6bb88f56905f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.686304] env[62600]: DEBUG oslo_concurrency.lockutils [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "5d7c2a17-9dee-407e-a7ee-6bb88f56905f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.686536] env[62600]: DEBUG oslo_concurrency.lockutils [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "5d7c2a17-9dee-407e-a7ee-6bb88f56905f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.686752] env[62600]: DEBUG oslo_concurrency.lockutils [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "5d7c2a17-9dee-407e-a7ee-6bb88f56905f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.688992] env[62600]: INFO nova.compute.manager [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Terminating instance [ 1022.690753] env[62600]: DEBUG nova.compute.manager [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1022.690957] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1022.691764] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d321aa2f-168b-408c-9e34-10e8a471ca67 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.698967] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1022.699199] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d9d7ff35-3ef2-442b-adb2-c3147d36b1ff {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.705915] env[62600]: DEBUG oslo_vmware.api [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1022.705915] env[62600]: value = "task-1222961" [ 1022.705915] env[62600]: _type = "Task" [ 1022.705915] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.715574] env[62600]: DEBUG oslo_vmware.api [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222961, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.716965] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1022.716965] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1022.717118] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Deleting the datastore file [datastore1] aff91190-e19b-4f1c-8b47-9539360a4596 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1022.717362] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a42a91ea-cfae-41bf-8d49-9b1a071061d8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.724735] env[62600]: DEBUG oslo_vmware.api [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for the task: (returnval){ [ 1022.724735] env[62600]: value = "task-1222962" [ 1022.724735] env[62600]: _type = "Task" [ 1022.724735] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.732926] env[62600]: DEBUG oslo_vmware.api [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222962, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.971123] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.698s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.973540] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 4.153s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.993774] env[62600]: INFO nova.scheduler.client.report [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Deleted allocations for instance ab11cdd1-80a7-43c4-afac-24c7930238c1 [ 1023.217600] env[62600]: DEBUG oslo_vmware.api [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222961, 'name': PowerOffVM_Task, 'duration_secs': 0.184026} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.218164] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1023.218164] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1023.218369] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-78bcb739-28da-4d27-9cb3-f26b1417b734 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.233205] env[62600]: DEBUG oslo_vmware.api [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Task: {'id': task-1222962, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142146} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.233422] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1023.233610] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1023.233790] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1023.233967] env[62600]: INFO nova.compute.manager [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1023.234219] env[62600]: DEBUG oslo.service.loopingcall [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1023.234410] env[62600]: DEBUG nova.compute.manager [-] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1023.234506] env[62600]: DEBUG nova.network.neutron [-] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1023.282827] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1023.282827] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1023.282827] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleting the datastore file [datastore1] 5d7c2a17-9dee-407e-a7ee-6bb88f56905f {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1023.283094] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-24917752-ce57-41b6-8cdb-192843950513 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.289134] env[62600]: DEBUG oslo_vmware.api [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1023.289134] env[62600]: value = "task-1222964" [ 1023.289134] env[62600]: _type = "Task" [ 1023.289134] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.299445] env[62600]: DEBUG oslo_vmware.api [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222964, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.478135] env[62600]: INFO nova.compute.claims [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1023.500206] env[62600]: DEBUG oslo_concurrency.lockutils [None req-8e4fa719-c4f6-4a4e-95bf-be62e1039195 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "ab11cdd1-80a7-43c4-afac-24c7930238c1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.350s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.509123] env[62600]: DEBUG nova.compute.manager [req-fe1a1da1-0e04-4330-b5ea-a037ccc83fac req-eb9ec19a-2c81-42dc-bbf5-be2216376f30 service nova] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Received event network-vif-deleted-7af8f843-5635-4c50-bc15-f8356b10e6f7 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1023.509323] env[62600]: INFO nova.compute.manager [req-fe1a1da1-0e04-4330-b5ea-a037ccc83fac req-eb9ec19a-2c81-42dc-bbf5-be2216376f30 service nova] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Neutron deleted interface 7af8f843-5635-4c50-bc15-f8356b10e6f7; detaching it from the instance and deleting it from the info cache [ 1023.509511] env[62600]: DEBUG nova.network.neutron [req-fe1a1da1-0e04-4330-b5ea-a037ccc83fac req-eb9ec19a-2c81-42dc-bbf5-be2216376f30 service nova] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.799325] env[62600]: DEBUG oslo_vmware.api [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222964, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.123483} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.799593] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1023.799786] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1023.799967] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1023.800162] env[62600]: INFO nova.compute.manager [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1023.800464] env[62600]: DEBUG oslo.service.loopingcall [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1023.800686] env[62600]: DEBUG nova.compute.manager [-] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1023.800784] env[62600]: DEBUG nova.network.neutron [-] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1023.986011] env[62600]: INFO nova.compute.resource_tracker [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Updating resource usage from migration a4591da5-9da8-4860-abec-9a3b0513c9a2 [ 1023.988761] env[62600]: DEBUG nova.network.neutron [-] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.012102] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cd2c54bc-4d42-4136-8e38-1fe75bec4c80 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.022018] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-276d9560-89fb-4d7f-ac64-18d17db3f639 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.052224] env[62600]: DEBUG nova.compute.manager [req-fe1a1da1-0e04-4330-b5ea-a037ccc83fac req-eb9ec19a-2c81-42dc-bbf5-be2216376f30 service nova] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Detach interface failed, port_id=7af8f843-5635-4c50-bc15-f8356b10e6f7, reason: Instance aff91190-e19b-4f1c-8b47-9539360a4596 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1024.153658] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-919cc670-628d-4c6d-9a58-8479cf637661 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.165307] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dee3e92-b25e-4e85-bffe-42b6befd518a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.200073] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a81cf65-78ca-4fa3-b3bc-287776117269 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.207633] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f15aa428-cd8d-44a3-bc59-dabfe0a31a6e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.223256] env[62600]: DEBUG nova.compute.provider_tree [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.492348] env[62600]: INFO nova.compute.manager [-] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Took 1.26 seconds to deallocate network for instance. [ 1024.591049] env[62600]: DEBUG nova.network.neutron [-] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.726407] env[62600]: DEBUG nova.scheduler.client.report [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1025.002054] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.093199] env[62600]: INFO nova.compute.manager [-] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Took 1.29 seconds to deallocate network for instance. [ 1025.231733] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.258s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.232026] env[62600]: INFO nova.compute.manager [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Migrating [ 1025.238465] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.175s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.240426] env[62600]: INFO nova.compute.claims [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1025.423185] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "8417575c-d68d-464f-8c22-ab358deb76d8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.423475] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "8417575c-d68d-464f-8c22-ab358deb76d8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.536835] env[62600]: DEBUG nova.compute.manager [req-a6a1f2ed-1d2e-4569-bb85-13e2c610f2a7 req-b7447ce6-f969-4ab2-b142-3b8b84cd5e5f service nova] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Received event network-vif-deleted-4c6c4d03-b2f6-4b73-acb7-7640e7047873 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1025.599625] env[62600]: DEBUG oslo_concurrency.lockutils [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.752291] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "refresh_cache-0416ef03-fd45-405f-9047-76b34a839217" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.752478] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "refresh_cache-0416ef03-fd45-405f-9047-76b34a839217" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.752659] env[62600]: DEBUG nova.network.neutron [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1025.925449] env[62600]: DEBUG nova.compute.manager [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1026.398445] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b32845d-4663-48a3-ab11-60e4799d75df {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.407724] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46c65a02-6794-47ab-aafb-e05230396fd2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.438996] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b4f6846-3a0e-4793-8743-59e52fe9b655 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.450156] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579456e7-f59b-4c52-a69e-da3bf5780902 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.455516] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.465195] env[62600]: DEBUG nova.compute.provider_tree [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.480564] env[62600]: DEBUG nova.network.neutron [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Updating instance_info_cache with network_info: [{"id": "bb15b28d-be8f-4fff-b8a6-0801d3e0541a", "address": "fa:16:3e:b1:43:47", "network": {"id": "95818650-8227-4d11-9d14-007f06c07790", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1738152115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73e664d178f7484a9f4741b4d9450e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb15b28d-be", "ovs_interfaceid": "bb15b28d-be8f-4fff-b8a6-0801d3e0541a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.970060] env[62600]: DEBUG nova.scheduler.client.report [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1026.983117] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "refresh_cache-0416ef03-fd45-405f-9047-76b34a839217" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.475273] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.237s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.475862] env[62600]: DEBUG nova.compute.manager [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1027.478739] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.954s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.478965] env[62600]: DEBUG nova.objects.instance [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lazy-loading 'resources' on Instance uuid 248a5371-6ff7-4da7-ae97-d638bea0123a {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.982134] env[62600]: DEBUG nova.compute.utils [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1027.986618] env[62600]: DEBUG nova.compute.manager [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1027.986804] env[62600]: DEBUG nova.network.neutron [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1028.026464] env[62600]: DEBUG nova.policy [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fbd1b2fb34d841359ada8fc44bec2986', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b028450f2da445fb83e37adfc86bba68', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 1028.116710] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c3ce310-b3fc-46f9-80e3-8381f070ecfe {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.125719] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84b2a94f-a21d-4af5-8faf-cf7c778bd309 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.155263] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-194d92a7-9dbd-4bad-9533-289e98ece8b4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.162851] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd20902-f8ed-4dec-b89f-ef968bc35ac2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.175381] env[62600]: DEBUG nova.compute.provider_tree [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.306310] env[62600]: DEBUG nova.network.neutron [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Successfully created port: b8e9cd2e-1990-4997-b1fa-cb4411d39fde {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1028.487274] env[62600]: DEBUG nova.compute.manager [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1028.498044] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb0f914-cb2e-4779-879d-bdf8551288cd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.519809] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Updating instance '0416ef03-fd45-405f-9047-76b34a839217' progress to 0 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1028.678490] env[62600]: DEBUG nova.scheduler.client.report [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1029.026552] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1029.027273] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e3f82f4-4149-49ab-a511-2b4b54ba7175 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.035860] env[62600]: DEBUG oslo_vmware.api [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1029.035860] env[62600]: value = "task-1222965" [ 1029.035860] env[62600]: _type = "Task" [ 1029.035860] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.044610] env[62600]: DEBUG oslo_vmware.api [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222965, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.183586] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.705s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.186333] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.184s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.186590] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.188598] env[62600]: DEBUG oslo_concurrency.lockutils [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.589s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.188878] env[62600]: DEBUG nova.objects.instance [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lazy-loading 'resources' on Instance uuid 5d7c2a17-9dee-407e-a7ee-6bb88f56905f {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.206063] env[62600]: INFO nova.scheduler.client.report [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Deleted allocations for instance aff91190-e19b-4f1c-8b47-9539360a4596 [ 1029.207823] env[62600]: INFO nova.scheduler.client.report [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Deleted allocations for instance 248a5371-6ff7-4da7-ae97-d638bea0123a [ 1029.497783] env[62600]: DEBUG nova.compute.manager [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1029.524065] env[62600]: DEBUG nova.virt.hardware [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1029.524340] env[62600]: DEBUG nova.virt.hardware [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1029.524508] env[62600]: DEBUG nova.virt.hardware [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1029.524726] env[62600]: DEBUG nova.virt.hardware [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1029.524883] env[62600]: DEBUG nova.virt.hardware [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1029.525052] env[62600]: DEBUG nova.virt.hardware [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1029.525277] env[62600]: DEBUG nova.virt.hardware [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1029.525516] env[62600]: DEBUG nova.virt.hardware [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1029.525726] env[62600]: DEBUG nova.virt.hardware [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1029.525902] env[62600]: DEBUG nova.virt.hardware [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1029.526100] env[62600]: DEBUG nova.virt.hardware [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1029.526981] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d4c187-c51f-4daf-9f16-e9c7e62177ed {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.535589] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c03fecf9-968c-4db1-a92c-c6c656bef474 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.546964] env[62600]: DEBUG oslo_vmware.api [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222965, 'name': PowerOffVM_Task, 'duration_secs': 0.182405} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.554483] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1029.554714] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Updating instance '0416ef03-fd45-405f-9047-76b34a839217' progress to 17 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1029.678535] env[62600]: DEBUG nova.compute.manager [req-0342c51b-8625-4916-b07a-af02e71f6d36 req-13e93728-4925-4003-8d86-0580a85fdf40 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Received event network-vif-plugged-b8e9cd2e-1990-4997-b1fa-cb4411d39fde {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1029.678901] env[62600]: DEBUG oslo_concurrency.lockutils [req-0342c51b-8625-4916-b07a-af02e71f6d36 req-13e93728-4925-4003-8d86-0580a85fdf40 service nova] Acquiring lock "eb3c5624-0599-4e81-b056-bedf34329646-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.679290] env[62600]: DEBUG oslo_concurrency.lockutils [req-0342c51b-8625-4916-b07a-af02e71f6d36 req-13e93728-4925-4003-8d86-0580a85fdf40 service nova] Lock "eb3c5624-0599-4e81-b056-bedf34329646-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.679593] env[62600]: DEBUG oslo_concurrency.lockutils [req-0342c51b-8625-4916-b07a-af02e71f6d36 req-13e93728-4925-4003-8d86-0580a85fdf40 service nova] Lock "eb3c5624-0599-4e81-b056-bedf34329646-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.679885] env[62600]: DEBUG nova.compute.manager [req-0342c51b-8625-4916-b07a-af02e71f6d36 req-13e93728-4925-4003-8d86-0580a85fdf40 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] No waiting events found dispatching network-vif-plugged-b8e9cd2e-1990-4997-b1fa-cb4411d39fde {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1029.680214] env[62600]: WARNING nova.compute.manager [req-0342c51b-8625-4916-b07a-af02e71f6d36 req-13e93728-4925-4003-8d86-0580a85fdf40 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Received unexpected event network-vif-plugged-b8e9cd2e-1990-4997-b1fa-cb4411d39fde for instance with vm_state building and task_state spawning. [ 1029.718113] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3923421d-10dd-4c7e-9b18-3d050c5e27f2 tempest-DeleteServersTestJSON-1539178256 tempest-DeleteServersTestJSON-1539178256-project-member] Lock "aff91190-e19b-4f1c-8b47-9539360a4596" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.611s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.719177] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3f96e7a6-52b8-4a19-89cc-a8036a5b55d3 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "248a5371-6ff7-4da7-ae97-d638bea0123a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.266s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.775452] env[62600]: DEBUG nova.network.neutron [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Successfully updated port: b8e9cd2e-1990-4997-b1fa-cb4411d39fde {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1029.839015] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-512f8694-b225-44d1-926d-55ab11fe8834 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.847134] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd08ab48-c2ff-4f28-97ee-64a121e5aa7a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.880142] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981cc761-3748-4dd2-808b-815615b447dd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.887457] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c27623-82a0-450c-9ab8-f93d1094e82e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.900571] env[62600]: DEBUG nova.compute.provider_tree [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.062400] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1030.062400] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1030.062400] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1030.062400] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1030.062400] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1030.062400] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1030.062708] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1030.063612] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1030.063612] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1030.063612] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1030.063612] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1030.069017] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-324e7a64-2daf-4c24-8889-f67aa5326737 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.087031] env[62600]: DEBUG oslo_vmware.api [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1030.087031] env[62600]: value = "task-1222966" [ 1030.087031] env[62600]: _type = "Task" [ 1030.087031] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.094417] env[62600]: DEBUG oslo_vmware.api [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222966, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.278343] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.278535] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.278744] env[62600]: DEBUG nova.network.neutron [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1030.407147] env[62600]: DEBUG nova.scheduler.client.report [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1030.596249] env[62600]: DEBUG oslo_vmware.api [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222966, 'name': ReconfigVM_Task, 'duration_secs': 0.159187} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.596616] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Updating instance '0416ef03-fd45-405f-9047-76b34a839217' progress to 33 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1030.905330] env[62600]: DEBUG nova.network.neutron [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1030.913788] env[62600]: DEBUG oslo_concurrency.lockutils [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.725s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.916737] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.461s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.917566] env[62600]: INFO nova.compute.claims [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1030.947801] env[62600]: INFO nova.scheduler.client.report [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleted allocations for instance 5d7c2a17-9dee-407e-a7ee-6bb88f56905f [ 1031.105235] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1031.105559] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1031.105744] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1031.105935] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1031.106097] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1031.106253] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1031.106543] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1031.106732] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1031.106908] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1031.107089] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1031.107276] env[62600]: DEBUG nova.virt.hardware [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1031.112814] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Reconfiguring VM instance instance-00000060 to detach disk 2000 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1031.113124] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-099321a6-044a-4f36-a6a2-9c3703ca92ed {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.129082] env[62600]: DEBUG nova.network.neutron [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Updating instance_info_cache with network_info: [{"id": "b8e9cd2e-1990-4997-b1fa-cb4411d39fde", "address": "fa:16:3e:fc:c1:ac", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8e9cd2e-19", "ovs_interfaceid": "b8e9cd2e-1990-4997-b1fa-cb4411d39fde", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.132141] env[62600]: DEBUG oslo_vmware.api [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1031.132141] env[62600]: value = "task-1222968" [ 1031.132141] env[62600]: _type = "Task" [ 1031.132141] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.141036] env[62600]: DEBUG oslo_vmware.api [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222968, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.455392] env[62600]: DEBUG oslo_concurrency.lockutils [None req-23c38ecc-ec14-4ea5-a95b-005ee3e608fe tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "5d7c2a17-9dee-407e-a7ee-6bb88f56905f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.769s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.631955] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.632305] env[62600]: DEBUG nova.compute.manager [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Instance network_info: |[{"id": "b8e9cd2e-1990-4997-b1fa-cb4411d39fde", "address": "fa:16:3e:fc:c1:ac", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8e9cd2e-19", "ovs_interfaceid": "b8e9cd2e-1990-4997-b1fa-cb4411d39fde", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1031.632861] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:c1:ac', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '184687d6-125a-4b58-bb5b-fdb404088eda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b8e9cd2e-1990-4997-b1fa-cb4411d39fde', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1031.641212] env[62600]: DEBUG oslo.service.loopingcall [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1031.644295] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1031.644520] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1ec59845-ead8-4cdb-8c42-8e9b2819d4d5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.664638] env[62600]: DEBUG oslo_vmware.api [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222968, 'name': ReconfigVM_Task, 'duration_secs': 0.161172} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.665333] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Reconfigured VM instance instance-00000060 to detach disk 2000 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1031.666108] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5cc76b9-0d97-4a21-b18f-b5884d82446a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.669009] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1031.669009] env[62600]: value = "task-1222969" [ 1031.669009] env[62600]: _type = "Task" [ 1031.669009] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.691842] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 0416ef03-fd45-405f-9047-76b34a839217/0416ef03-fd45-405f-9047-76b34a839217.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1031.692582] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-764d67d9-2d16-40d4-9d98-19ae1bd5af8d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.712708] env[62600]: DEBUG nova.compute.manager [req-4799668e-cd07-4f9b-a20e-aee5dfb04d60 req-4a0d61b6-c36b-47d1-bbc1-ef919e7a3f12 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Received event network-changed-b8e9cd2e-1990-4997-b1fa-cb4411d39fde {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1031.712910] env[62600]: DEBUG nova.compute.manager [req-4799668e-cd07-4f9b-a20e-aee5dfb04d60 req-4a0d61b6-c36b-47d1-bbc1-ef919e7a3f12 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Refreshing instance network info cache due to event network-changed-b8e9cd2e-1990-4997-b1fa-cb4411d39fde. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1031.713142] env[62600]: DEBUG oslo_concurrency.lockutils [req-4799668e-cd07-4f9b-a20e-aee5dfb04d60 req-4a0d61b6-c36b-47d1-bbc1-ef919e7a3f12 service nova] Acquiring lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.713291] env[62600]: DEBUG oslo_concurrency.lockutils [req-4799668e-cd07-4f9b-a20e-aee5dfb04d60 req-4a0d61b6-c36b-47d1-bbc1-ef919e7a3f12 service nova] Acquired lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.713464] env[62600]: DEBUG nova.network.neutron [req-4799668e-cd07-4f9b-a20e-aee5dfb04d60 req-4a0d61b6-c36b-47d1-bbc1-ef919e7a3f12 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Refreshing network info cache for port b8e9cd2e-1990-4997-b1fa-cb4411d39fde {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1031.714723] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222969, 'name': CreateVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.719892] env[62600]: DEBUG oslo_vmware.api [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1031.719892] env[62600]: value = "task-1222970" [ 1031.719892] env[62600]: _type = "Task" [ 1031.719892] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.728556] env[62600]: DEBUG oslo_vmware.api [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222970, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.055011] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b463cd1c-6880-47d0-a515-9c2fbe2b6203 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.064899] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9a663a2-58e3-44e5-a537-6bc6d5c13fe0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.096290] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b416cb1e-22a8-489c-ab73-e56db98dbf85 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.103765] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-632e8f31-e460-4032-a34e-8ff4ba84421a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.117684] env[62600]: DEBUG nova.compute.provider_tree [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.184382] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222969, 'name': CreateVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.228673] env[62600]: DEBUG oslo_vmware.api [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222970, 'name': ReconfigVM_Task, 'duration_secs': 0.24635} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.228968] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 0416ef03-fd45-405f-9047-76b34a839217/0416ef03-fd45-405f-9047-76b34a839217.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1032.230445] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Updating instance '0416ef03-fd45-405f-9047-76b34a839217' progress to 50 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1032.455610] env[62600]: DEBUG nova.network.neutron [req-4799668e-cd07-4f9b-a20e-aee5dfb04d60 req-4a0d61b6-c36b-47d1-bbc1-ef919e7a3f12 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Updated VIF entry in instance network info cache for port b8e9cd2e-1990-4997-b1fa-cb4411d39fde. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1032.455610] env[62600]: DEBUG nova.network.neutron [req-4799668e-cd07-4f9b-a20e-aee5dfb04d60 req-4a0d61b6-c36b-47d1-bbc1-ef919e7a3f12 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Updating instance_info_cache with network_info: [{"id": "b8e9cd2e-1990-4997-b1fa-cb4411d39fde", "address": "fa:16:3e:fc:c1:ac", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8e9cd2e-19", "ovs_interfaceid": "b8e9cd2e-1990-4997-b1fa-cb4411d39fde", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.620267] env[62600]: DEBUG nova.scheduler.client.report [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1032.680221] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222969, 'name': CreateVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.739068] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8540be0-26d9-4b42-aa3a-85c1c79a787f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.758355] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9496988-0b8f-4b63-b47b-cdefdfd6f679 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.777235] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Updating instance '0416ef03-fd45-405f-9047-76b34a839217' progress to 67 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1032.957138] env[62600]: DEBUG oslo_concurrency.lockutils [req-4799668e-cd07-4f9b-a20e-aee5dfb04d60 req-4a0d61b6-c36b-47d1-bbc1-ef919e7a3f12 service nova] Releasing lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.000785] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "4607dacd-28ab-4173-ba86-8a78bec19dbc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.001078] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "4607dacd-28ab-4173-ba86-8a78bec19dbc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.125650] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.209s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.126208] env[62600]: DEBUG nova.compute.manager [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1033.179888] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222969, 'name': CreateVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.343113] env[62600]: DEBUG nova.network.neutron [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Port bb15b28d-be8f-4fff-b8a6-0801d3e0541a binding to destination host cpu-1 is already ACTIVE {{(pid=62600) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1033.503830] env[62600]: DEBUG nova.compute.manager [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1033.631628] env[62600]: DEBUG nova.compute.utils [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1033.633035] env[62600]: DEBUG nova.compute.manager [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1033.633202] env[62600]: DEBUG nova.network.neutron [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1033.676149] env[62600]: DEBUG nova.policy [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '94751040d0fd41a895010c0f2f019131', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef3bb131f3994f7a8673dafff6d353c1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 1033.683344] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222969, 'name': CreateVM_Task, 'duration_secs': 1.842598} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.683651] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1033.684173] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.684497] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.684702] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1033.684901] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d79b7e7-aa7f-4b0f-8288-17e53ee523a5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.689198] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1033.689198] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520421c4-2f6d-685d-fe38-df663fce3682" [ 1033.689198] env[62600]: _type = "Task" [ 1033.689198] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.696601] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520421c4-2f6d-685d-fe38-df663fce3682, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.954828] env[62600]: DEBUG nova.network.neutron [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Successfully created port: 028950a6-9715-4e67-b31b-447158a62875 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1034.028860] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.029462] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.031998] env[62600]: INFO nova.compute.claims [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1034.136932] env[62600]: DEBUG nova.compute.manager [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1034.199958] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520421c4-2f6d-685d-fe38-df663fce3682, 'name': SearchDatastore_Task, 'duration_secs': 0.009895} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.200387] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.200699] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1034.201011] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.201245] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.201500] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1034.201819] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-129e1dc6-dccf-4467-820d-57c16f487bfa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.210275] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1034.210586] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1034.211320] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8550799b-7ca6-4380-b4d5-cee33067c216 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.216391] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1034.216391] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]522f364b-3a24-753c-bb2b-42c3a8aa7a27" [ 1034.216391] env[62600]: _type = "Task" [ 1034.216391] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.223972] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]522f364b-3a24-753c-bb2b-42c3a8aa7a27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.364824] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "0416ef03-fd45-405f-9047-76b34a839217-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.365087] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "0416ef03-fd45-405f-9047-76b34a839217-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.365272] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "0416ef03-fd45-405f-9047-76b34a839217-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.729307] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]522f364b-3a24-753c-bb2b-42c3a8aa7a27, 'name': SearchDatastore_Task, 'duration_secs': 0.008706} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.730326] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-926c6d0e-d9b0-478c-a983-4c44fe61caac {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.736293] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1034.736293] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b2b42a-4fb6-32d3-f8a3-e1885f333a20" [ 1034.736293] env[62600]: _type = "Task" [ 1034.736293] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.744826] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b2b42a-4fb6-32d3-f8a3-e1885f333a20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.878600] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "d66c404d-8fd1-4fb7-a3b9-f21854c7e735" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.879798] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "d66c404d-8fd1-4fb7-a3b9-f21854c7e735" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.879798] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "d66c404d-8fd1-4fb7-a3b9-f21854c7e735-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.879798] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "d66c404d-8fd1-4fb7-a3b9-f21854c7e735-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.879798] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "d66c404d-8fd1-4fb7-a3b9-f21854c7e735-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.883230] env[62600]: INFO nova.compute.manager [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Terminating instance [ 1034.885353] env[62600]: DEBUG nova.compute.manager [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1034.885559] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1034.886458] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f746fb96-6d8d-4da7-b58d-4b314cdd3ef7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.895139] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1034.896757] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6a19b7f8-836c-4894-b43e-884e34eacca8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.903035] env[62600]: DEBUG oslo_vmware.api [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 1034.903035] env[62600]: value = "task-1222971" [ 1034.903035] env[62600]: _type = "Task" [ 1034.903035] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.912220] env[62600]: DEBUG oslo_vmware.api [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222971, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.151050] env[62600]: DEBUG nova.compute.manager [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1035.177695] env[62600]: DEBUG nova.virt.hardware [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1035.178406] env[62600]: DEBUG nova.virt.hardware [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1035.178406] env[62600]: DEBUG nova.virt.hardware [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1035.178406] env[62600]: DEBUG nova.virt.hardware [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1035.178541] env[62600]: DEBUG nova.virt.hardware [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1035.178683] env[62600]: DEBUG nova.virt.hardware [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1035.178902] env[62600]: DEBUG nova.virt.hardware [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1035.179097] env[62600]: DEBUG nova.virt.hardware [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1035.179293] env[62600]: DEBUG nova.virt.hardware [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1035.179497] env[62600]: DEBUG nova.virt.hardware [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1035.179700] env[62600]: DEBUG nova.virt.hardware [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1035.180595] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cda2554-8401-405e-a5df-85c5682b33af {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.184560] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c56873e-4e50-4efc-9b6a-7cc580f64836 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.195159] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd4033c6-9368-4c9b-b57d-ce71accdc0cf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.201246] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f60cb79-266b-402d-a71d-3790a095fce1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.237933] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89b9c1fe-5395-478b-8b14-ca6a6dda6dd6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.249878] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52b2b42a-4fb6-32d3-f8a3-e1885f333a20, 'name': SearchDatastore_Task, 'duration_secs': 0.00967} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.252362] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.252649] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] eb3c5624-0599-4e81-b056-bedf34329646/eb3c5624-0599-4e81-b056-bedf34329646.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1035.253135] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0ad91ad0-2c43-45fc-8ede-55b049e051ab {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.259412] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a172689e-3043-41d5-8e58-94a4756ba4b5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.270431] env[62600]: DEBUG nova.compute.provider_tree [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.272857] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1035.272857] env[62600]: value = "task-1222972" [ 1035.272857] env[62600]: _type = "Task" [ 1035.272857] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.281583] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222972, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.381289] env[62600]: DEBUG nova.compute.manager [req-50104d89-8861-49cb-adb0-aed380a7f7a8 req-ef2708ef-3f9c-4e5e-b08d-5aa84d2d4adf service nova] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Received event network-vif-plugged-028950a6-9715-4e67-b31b-447158a62875 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1035.381545] env[62600]: DEBUG oslo_concurrency.lockutils [req-50104d89-8861-49cb-adb0-aed380a7f7a8 req-ef2708ef-3f9c-4e5e-b08d-5aa84d2d4adf service nova] Acquiring lock "8417575c-d68d-464f-8c22-ab358deb76d8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.381776] env[62600]: DEBUG oslo_concurrency.lockutils [req-50104d89-8861-49cb-adb0-aed380a7f7a8 req-ef2708ef-3f9c-4e5e-b08d-5aa84d2d4adf service nova] Lock "8417575c-d68d-464f-8c22-ab358deb76d8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.381969] env[62600]: DEBUG oslo_concurrency.lockutils [req-50104d89-8861-49cb-adb0-aed380a7f7a8 req-ef2708ef-3f9c-4e5e-b08d-5aa84d2d4adf service nova] Lock "8417575c-d68d-464f-8c22-ab358deb76d8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.382156] env[62600]: DEBUG nova.compute.manager [req-50104d89-8861-49cb-adb0-aed380a7f7a8 req-ef2708ef-3f9c-4e5e-b08d-5aa84d2d4adf service nova] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] No waiting events found dispatching network-vif-plugged-028950a6-9715-4e67-b31b-447158a62875 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1035.382329] env[62600]: WARNING nova.compute.manager [req-50104d89-8861-49cb-adb0-aed380a7f7a8 req-ef2708ef-3f9c-4e5e-b08d-5aa84d2d4adf service nova] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Received unexpected event network-vif-plugged-028950a6-9715-4e67-b31b-447158a62875 for instance with vm_state building and task_state spawning. [ 1035.418592] env[62600]: DEBUG oslo_vmware.api [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222971, 'name': PowerOffVM_Task, 'duration_secs': 0.224279} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.419903] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "refresh_cache-0416ef03-fd45-405f-9047-76b34a839217" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.420143] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "refresh_cache-0416ef03-fd45-405f-9047-76b34a839217" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.420331] env[62600]: DEBUG nova.network.neutron [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1035.421977] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1035.422142] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1035.422433] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b024215c-023a-46f9-8257-668154ab5a8e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.462289] env[62600]: DEBUG nova.network.neutron [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Successfully updated port: 028950a6-9715-4e67-b31b-447158a62875 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1035.511615] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1035.511970] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1035.512209] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Deleting the datastore file [datastore1] d66c404d-8fd1-4fb7-a3b9-f21854c7e735 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1035.512513] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d1e8080-5156-4a33-ac1f-4914e153d28c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.520769] env[62600]: DEBUG oslo_vmware.api [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 1035.520769] env[62600]: value = "task-1222974" [ 1035.520769] env[62600]: _type = "Task" [ 1035.520769] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.529238] env[62600]: DEBUG oslo_vmware.api [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222974, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.775165] env[62600]: DEBUG nova.scheduler.client.report [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1035.788889] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222972, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.452675} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.789251] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] eb3c5624-0599-4e81-b056-bedf34329646/eb3c5624-0599-4e81-b056-bedf34329646.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1035.789713] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1035.789780] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ccc49e71-58c1-4ed8-886f-2e2f89d4e57f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.796817] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1035.796817] env[62600]: value = "task-1222975" [ 1035.796817] env[62600]: _type = "Task" [ 1035.796817] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.805335] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222975, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.966222] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "refresh_cache-8417575c-d68d-464f-8c22-ab358deb76d8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.966404] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquired lock "refresh_cache-8417575c-d68d-464f-8c22-ab358deb76d8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.966565] env[62600]: DEBUG nova.network.neutron [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1036.035155] env[62600]: DEBUG oslo_vmware.api [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222974, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.359824} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.035155] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1036.035155] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1036.035155] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1036.035155] env[62600]: INFO nova.compute.manager [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1036.035155] env[62600]: DEBUG oslo.service.loopingcall [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.035155] env[62600]: DEBUG nova.compute.manager [-] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1036.035155] env[62600]: DEBUG nova.network.neutron [-] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1036.141607] env[62600]: DEBUG nova.network.neutron [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Updating instance_info_cache with network_info: [{"id": "bb15b28d-be8f-4fff-b8a6-0801d3e0541a", "address": "fa:16:3e:b1:43:47", "network": {"id": "95818650-8227-4d11-9d14-007f06c07790", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1738152115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73e664d178f7484a9f4741b4d9450e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb15b28d-be", "ovs_interfaceid": "bb15b28d-be8f-4fff-b8a6-0801d3e0541a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.282178] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.251s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.282178] env[62600]: DEBUG nova.compute.manager [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1036.296317] env[62600]: DEBUG nova.compute.manager [req-ac599a38-11c1-44dc-9134-c320df69c196 req-b4b4970a-8b5e-4575-af23-06379328bf83 service nova] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Received event network-vif-deleted-57606b61-13bc-4562-b28d-d61f215809e0 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1036.296317] env[62600]: INFO nova.compute.manager [req-ac599a38-11c1-44dc-9134-c320df69c196 req-b4b4970a-8b5e-4575-af23-06379328bf83 service nova] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Neutron deleted interface 57606b61-13bc-4562-b28d-d61f215809e0; detaching it from the instance and deleting it from the info cache [ 1036.296317] env[62600]: DEBUG nova.network.neutron [req-ac599a38-11c1-44dc-9134-c320df69c196 req-b4b4970a-8b5e-4575-af23-06379328bf83 service nova] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.309943] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222975, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060406} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.310246] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1036.311048] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77907dc9-d4fa-440e-ad59-de0f8d4170eb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.340388] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] eb3c5624-0599-4e81-b056-bedf34329646/eb3c5624-0599-4e81-b056-bedf34329646.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1036.340963] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2135286b-3842-4570-bc25-615b97376ab1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.367644] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1036.367644] env[62600]: value = "task-1222976" [ 1036.367644] env[62600]: _type = "Task" [ 1036.367644] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.376749] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222976, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.497128] env[62600]: DEBUG nova.network.neutron [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1036.578512] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "692fa608-39f4-4d7d-9f30-b6559878b970" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.578752] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "692fa608-39f4-4d7d-9f30-b6559878b970" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.628352] env[62600]: DEBUG nova.network.neutron [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Updating instance_info_cache with network_info: [{"id": "028950a6-9715-4e67-b31b-447158a62875", "address": "fa:16:3e:f9:ef:88", "network": {"id": "e7547ac4-24a6-4778-a17a-40e79e7a0d0c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-320671681-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3bb131f3994f7a8673dafff6d353c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap028950a6-97", "ovs_interfaceid": "028950a6-9715-4e67-b31b-447158a62875", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.644479] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "refresh_cache-0416ef03-fd45-405f-9047-76b34a839217" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.728012] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1036.770266] env[62600]: DEBUG nova.network.neutron [-] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.788582] env[62600]: DEBUG nova.compute.utils [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1036.789936] env[62600]: DEBUG nova.compute.manager [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1036.790125] env[62600]: DEBUG nova.network.neutron [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1036.800528] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-32db4654-6eed-4602-add8-ff9a4f08987c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.810063] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cfb7772-9e5f-472c-bfd9-7fd7cbcce669 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.835964] env[62600]: DEBUG nova.compute.manager [req-ac599a38-11c1-44dc-9134-c320df69c196 req-b4b4970a-8b5e-4575-af23-06379328bf83 service nova] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Detach interface failed, port_id=57606b61-13bc-4562-b28d-d61f215809e0, reason: Instance d66c404d-8fd1-4fb7-a3b9-f21854c7e735 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1036.873008] env[62600]: DEBUG nova.policy [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3e59998217a4b18a6f3f01142a5e440', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f91091f83ee4a2091507ca994e3d52f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 1036.880683] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222976, 'name': ReconfigVM_Task, 'duration_secs': 0.331915} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.881122] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Reconfigured VM instance instance-00000061 to attach disk [datastore1] eb3c5624-0599-4e81-b056-bedf34329646/eb3c5624-0599-4e81-b056-bedf34329646.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1036.882145] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cbf71a38-d200-4a44-95b5-ea4c7711a5d3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.888653] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1036.888653] env[62600]: value = "task-1222977" [ 1036.888653] env[62600]: _type = "Task" [ 1036.888653] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.897033] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222977, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.083622] env[62600]: DEBUG nova.compute.manager [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1037.133015] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Releasing lock "refresh_cache-8417575c-d68d-464f-8c22-ab358deb76d8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.133015] env[62600]: DEBUG nova.compute.manager [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Instance network_info: |[{"id": "028950a6-9715-4e67-b31b-447158a62875", "address": "fa:16:3e:f9:ef:88", "network": {"id": "e7547ac4-24a6-4778-a17a-40e79e7a0d0c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-320671681-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3bb131f3994f7a8673dafff6d353c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap028950a6-97", "ovs_interfaceid": "028950a6-9715-4e67-b31b-447158a62875", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1037.133015] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f9:ef:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '028950a6-9715-4e67-b31b-447158a62875', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1037.141202] env[62600]: DEBUG oslo.service.loopingcall [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1037.141827] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1037.144986] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-38d05eed-0b0c-44df-a16c-c3f37e1ea7f7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.161991] env[62600]: DEBUG nova.network.neutron [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Successfully created port: acb612e6-f57e-4eae-8764-c9525e4c6baa {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1037.175105] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1037.175105] env[62600]: value = "task-1222978" [ 1037.175105] env[62600]: _type = "Task" [ 1037.175105] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.180774] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222978, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.189018] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f63654a9-f528-45c9-8aae-a8a8892600a7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.206766] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c6d937f-91e2-4927-bd2b-63e4d3570fb0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.214324] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Updating instance '0416ef03-fd45-405f-9047-76b34a839217' progress to 83 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1037.230822] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.233017] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.233017] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.233017] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62600) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1037.233017] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fda6e6f-4fea-45df-8dc4-b01223166383 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.240513] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-983149a7-eb84-4717-82af-000479797e8b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.260718] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9a0bb95-3f5a-4a37-8832-e4cf2726a0d2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.269253] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d44e6836-d401-4fcc-9a7f-b72b42424d66 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.275456] env[62600]: INFO nova.compute.manager [-] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Took 1.24 seconds to deallocate network for instance. [ 1037.302023] env[62600]: DEBUG nova.compute.manager [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1037.305095] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179628MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=62600) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1037.305095] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.305095] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.399386] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222977, 'name': Rename_Task, 'duration_secs': 0.215408} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.399717] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1037.399972] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-61ffa1f0-639a-490b-9238-afc82e8004ab {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.406782] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1037.406782] env[62600]: value = "task-1222979" [ 1037.406782] env[62600]: _type = "Task" [ 1037.406782] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.414595] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222979, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.422609] env[62600]: DEBUG nova.compute.manager [req-6802e646-5112-4243-aa1a-59c3b208c68c req-210429b9-2e77-48ee-b069-4eacdb364ece service nova] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Received event network-changed-028950a6-9715-4e67-b31b-447158a62875 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1037.423057] env[62600]: DEBUG nova.compute.manager [req-6802e646-5112-4243-aa1a-59c3b208c68c req-210429b9-2e77-48ee-b069-4eacdb364ece service nova] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Refreshing instance network info cache due to event network-changed-028950a6-9715-4e67-b31b-447158a62875. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1037.423511] env[62600]: DEBUG oslo_concurrency.lockutils [req-6802e646-5112-4243-aa1a-59c3b208c68c req-210429b9-2e77-48ee-b069-4eacdb364ece service nova] Acquiring lock "refresh_cache-8417575c-d68d-464f-8c22-ab358deb76d8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.423614] env[62600]: DEBUG oslo_concurrency.lockutils [req-6802e646-5112-4243-aa1a-59c3b208c68c req-210429b9-2e77-48ee-b069-4eacdb364ece service nova] Acquired lock "refresh_cache-8417575c-d68d-464f-8c22-ab358deb76d8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.423817] env[62600]: DEBUG nova.network.neutron [req-6802e646-5112-4243-aa1a-59c3b208c68c req-210429b9-2e77-48ee-b069-4eacdb364ece service nova] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Refreshing network info cache for port 028950a6-9715-4e67-b31b-447158a62875 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1037.599529] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.682756] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222978, 'name': CreateVM_Task, 'duration_secs': 0.342032} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.682922] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1037.683678] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.683887] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.684246] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1037.684842] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c469744d-dfb6-4003-8e65-fc096d11fa51 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.689471] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1037.689471] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5275bbe8-6e95-c077-e5bf-e6846cf4e81a" [ 1037.689471] env[62600]: _type = "Task" [ 1037.689471] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.697370] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5275bbe8-6e95-c077-e5bf-e6846cf4e81a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.721238] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1037.721510] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f2710ca-4d2a-4c8f-828f-3b71c0eb9752 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.728193] env[62600]: DEBUG oslo_vmware.api [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1037.728193] env[62600]: value = "task-1222980" [ 1037.728193] env[62600]: _type = "Task" [ 1037.728193] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.737143] env[62600]: DEBUG oslo_vmware.api [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222980, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.816171] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.917773] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222979, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.157505] env[62600]: DEBUG nova.network.neutron [req-6802e646-5112-4243-aa1a-59c3b208c68c req-210429b9-2e77-48ee-b069-4eacdb364ece service nova] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Updated VIF entry in instance network info cache for port 028950a6-9715-4e67-b31b-447158a62875. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1038.158012] env[62600]: DEBUG nova.network.neutron [req-6802e646-5112-4243-aa1a-59c3b208c68c req-210429b9-2e77-48ee-b069-4eacdb364ece service nova] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Updating instance_info_cache with network_info: [{"id": "028950a6-9715-4e67-b31b-447158a62875", "address": "fa:16:3e:f9:ef:88", "network": {"id": "e7547ac4-24a6-4778-a17a-40e79e7a0d0c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-320671681-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3bb131f3994f7a8673dafff6d353c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap028950a6-97", "ovs_interfaceid": "028950a6-9715-4e67-b31b-447158a62875", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.200157] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5275bbe8-6e95-c077-e5bf-e6846cf4e81a, 'name': SearchDatastore_Task, 'duration_secs': 0.015443} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.200449] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.200694] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1038.200930] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.201127] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.201335] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1038.201602] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9ecef5f2-33ac-4a61-8e54-c99b61716d70 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.214332] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1038.214521] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1038.215231] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77c240c0-425d-4e42-aa3d-1f0c35c71344 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.221137] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1038.221137] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f856e2-5031-870f-0d1d-6e5440aca5cd" [ 1038.221137] env[62600]: _type = "Task" [ 1038.221137] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.229032] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f856e2-5031-870f-0d1d-6e5440aca5cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.237260] env[62600]: DEBUG oslo_vmware.api [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222980, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.318182] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Applying migration context for instance 0416ef03-fd45-405f-9047-76b34a839217 as it has an incoming, in-progress migration a4591da5-9da8-4860-abec-9a3b0513c9a2. Migration status is post-migrating {{(pid=62600) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1038.319561] env[62600]: INFO nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Updating resource usage from migration a4591da5-9da8-4860-abec-9a3b0513c9a2 [ 1038.322735] env[62600]: DEBUG nova.compute.manager [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1038.344684] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance b967fb11-e70e-4e17-b769-38da581bd83b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1038.344872] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance d66c404d-8fd1-4fb7-a3b9-f21854c7e735 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1038.345055] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 1b856aff-66be-4fa5-b8b7-124ac24fb4d1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1038.345193] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance c2f38d56-72fa-4bf0-a2cb-13d666571466 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1038.345345] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Migration a4591da5-9da8-4860-abec-9a3b0513c9a2 is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1038.345459] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 0416ef03-fd45-405f-9047-76b34a839217 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1038.345572] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance eb3c5624-0599-4e81-b056-bedf34329646 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1038.345685] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 8417575c-d68d-464f-8c22-ab358deb76d8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1038.345796] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 4607dacd-28ab-4173-ba86-8a78bec19dbc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1038.349364] env[62600]: DEBUG nova.virt.hardware [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1038.349618] env[62600]: DEBUG nova.virt.hardware [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1038.349782] env[62600]: DEBUG nova.virt.hardware [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1038.349972] env[62600]: DEBUG nova.virt.hardware [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1038.350143] env[62600]: DEBUG nova.virt.hardware [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1038.350298] env[62600]: DEBUG nova.virt.hardware [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1038.350511] env[62600]: DEBUG nova.virt.hardware [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1038.350678] env[62600]: DEBUG nova.virt.hardware [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1038.350850] env[62600]: DEBUG nova.virt.hardware [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1038.351025] env[62600]: DEBUG nova.virt.hardware [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1038.351210] env[62600]: DEBUG nova.virt.hardware [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1038.352088] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc8234d-45cc-4d7f-afe3-ce0badb72037 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.357380] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 692fa608-39f4-4d7d-9f30-b6559878b970 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1038.357590] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1038.357732] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2304MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1038.365276] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a78fedd-5d99-40be-a1a7-81c3d7de82d2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.422671] env[62600]: DEBUG oslo_vmware.api [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1222979, 'name': PowerOnVM_Task, 'duration_secs': 0.540888} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.423081] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1038.423389] env[62600]: INFO nova.compute.manager [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Took 8.93 seconds to spawn the instance on the hypervisor. [ 1038.423668] env[62600]: DEBUG nova.compute.manager [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1038.424781] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c2de19e-0056-4d32-a323-f47c3b86c886 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.497097] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6d1cc7-e573-480a-b436-f4edf663c69d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.504043] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ad9150-fcea-4c2c-9557-ea9c070c8919 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.533401] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0380ffe3-e952-47b6-9810-7c4b60a831a0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.540604] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ddde212-d3f6-4cc4-a1d2-7ee904db8b84 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.553312] env[62600]: DEBUG nova.compute.provider_tree [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1038.660314] env[62600]: DEBUG oslo_concurrency.lockutils [req-6802e646-5112-4243-aa1a-59c3b208c68c req-210429b9-2e77-48ee-b069-4eacdb364ece service nova] Releasing lock "refresh_cache-8417575c-d68d-464f-8c22-ab358deb76d8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.685518] env[62600]: DEBUG nova.compute.manager [req-9d5026e6-1f23-441b-8a2f-28ab0138e1db req-c1603b93-7096-4295-9b4d-66a0042f1ba3 service nova] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Received event network-vif-plugged-acb612e6-f57e-4eae-8764-c9525e4c6baa {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1038.685752] env[62600]: DEBUG oslo_concurrency.lockutils [req-9d5026e6-1f23-441b-8a2f-28ab0138e1db req-c1603b93-7096-4295-9b4d-66a0042f1ba3 service nova] Acquiring lock "4607dacd-28ab-4173-ba86-8a78bec19dbc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.685969] env[62600]: DEBUG oslo_concurrency.lockutils [req-9d5026e6-1f23-441b-8a2f-28ab0138e1db req-c1603b93-7096-4295-9b4d-66a0042f1ba3 service nova] Lock "4607dacd-28ab-4173-ba86-8a78bec19dbc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.686162] env[62600]: DEBUG oslo_concurrency.lockutils [req-9d5026e6-1f23-441b-8a2f-28ab0138e1db req-c1603b93-7096-4295-9b4d-66a0042f1ba3 service nova] Lock "4607dacd-28ab-4173-ba86-8a78bec19dbc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.686358] env[62600]: DEBUG nova.compute.manager [req-9d5026e6-1f23-441b-8a2f-28ab0138e1db req-c1603b93-7096-4295-9b4d-66a0042f1ba3 service nova] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] No waiting events found dispatching network-vif-plugged-acb612e6-f57e-4eae-8764-c9525e4c6baa {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1038.686548] env[62600]: WARNING nova.compute.manager [req-9d5026e6-1f23-441b-8a2f-28ab0138e1db req-c1603b93-7096-4295-9b4d-66a0042f1ba3 service nova] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Received unexpected event network-vif-plugged-acb612e6-f57e-4eae-8764-c9525e4c6baa for instance with vm_state building and task_state spawning. [ 1038.735335] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f856e2-5031-870f-0d1d-6e5440aca5cd, 'name': SearchDatastore_Task, 'duration_secs': 0.028466} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.740786] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e1c5567-e5d3-4dcb-8355-dfa8f689daf1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.747579] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1038.747579] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52404da6-7ba1-cec0-f36c-a8cdedd50bb2" [ 1038.747579] env[62600]: _type = "Task" [ 1038.747579] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.751370] env[62600]: DEBUG oslo_vmware.api [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1222980, 'name': PowerOnVM_Task, 'duration_secs': 0.626787} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.756124] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1038.756486] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a787d84b-4bb6-49fa-9535-7550b78d8269 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Updating instance '0416ef03-fd45-405f-9047-76b34a839217' progress to 100 {{(pid=62600) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1038.772925] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52404da6-7ba1-cec0-f36c-a8cdedd50bb2, 'name': SearchDatastore_Task, 'duration_secs': 0.015545} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.773563] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.773839] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 8417575c-d68d-464f-8c22-ab358deb76d8/8417575c-d68d-464f-8c22-ab358deb76d8.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1038.774121] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d63ae3c3-380c-4c5d-9034-8e8707064f41 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.781104] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1038.781104] env[62600]: value = "task-1222981" [ 1038.781104] env[62600]: _type = "Task" [ 1038.781104] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.789829] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222981, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.860747] env[62600]: DEBUG nova.network.neutron [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Successfully updated port: acb612e6-f57e-4eae-8764-c9525e4c6baa {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1038.946675] env[62600]: INFO nova.compute.manager [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Took 19.90 seconds to build instance. [ 1039.059276] env[62600]: DEBUG nova.scheduler.client.report [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1039.294114] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222981, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.363550] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "refresh_cache-4607dacd-28ab-4173-ba86-8a78bec19dbc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.363743] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "refresh_cache-4607dacd-28ab-4173-ba86-8a78bec19dbc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.363935] env[62600]: DEBUG nova.network.neutron [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1039.449516] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e87c35ea-bf8f-491c-b2f6-6d0b84caa150 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "eb3c5624-0599-4e81-b056-bedf34329646" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.406s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.567025] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62600) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1039.567025] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.260s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.567025] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.966s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.567274] env[62600]: INFO nova.compute.claims [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1039.791478] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222981, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534334} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.791738] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 8417575c-d68d-464f-8c22-ab358deb76d8/8417575c-d68d-464f-8c22-ab358deb76d8.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1039.791954] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1039.792230] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d70255bd-4d31-4d43-9357-b00d9080648d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.798981] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1039.798981] env[62600]: value = "task-1222982" [ 1039.798981] env[62600]: _type = "Task" [ 1039.798981] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.811468] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222982, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.904155] env[62600]: DEBUG nova.network.neutron [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1040.060386] env[62600]: DEBUG nova.network.neutron [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Updating instance_info_cache with network_info: [{"id": "acb612e6-f57e-4eae-8764-c9525e4c6baa", "address": "fa:16:3e:ef:40:fe", "network": {"id": "1d2052cf-5be1-48e7-887a-18f367b7823f", "bridge": "br-int", "label": "tempest-ServersTestJSON-106099679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f91091f83ee4a2091507ca994e3d52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapacb612e6-f5", "ovs_interfaceid": "acb612e6-f57e-4eae-8764-c9525e4c6baa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.311038] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222982, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063332} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.311356] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1040.312197] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a443b7-1aab-47f3-832f-b19cae61a073 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.338070] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 8417575c-d68d-464f-8c22-ab358deb76d8/8417575c-d68d-464f-8c22-ab358deb76d8.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1040.338070] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6d1db66-1644-4efe-9e9d-4bcedd36883e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.359751] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1040.359751] env[62600]: value = "task-1222983" [ 1040.359751] env[62600]: _type = "Task" [ 1040.359751] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.367953] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222983, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.563737] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "refresh_cache-4607dacd-28ab-4173-ba86-8a78bec19dbc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.563876] env[62600]: DEBUG nova.compute.manager [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Instance network_info: |[{"id": "acb612e6-f57e-4eae-8764-c9525e4c6baa", "address": "fa:16:3e:ef:40:fe", "network": {"id": "1d2052cf-5be1-48e7-887a-18f367b7823f", "bridge": "br-int", "label": "tempest-ServersTestJSON-106099679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f91091f83ee4a2091507ca994e3d52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapacb612e6-f5", "ovs_interfaceid": "acb612e6-f57e-4eae-8764-c9525e4c6baa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1040.564356] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ef:40:fe', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '028bae2d-fe6c-4207-b4a3-3fab45fbf1d6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'acb612e6-f57e-4eae-8764-c9525e4c6baa', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1040.572543] env[62600]: DEBUG oslo.service.loopingcall [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1040.572927] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1040.573167] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1040.573427] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1040.573623] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-822a5377-ed2e-4502-8c87-22ab7612f9dc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.588716] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1040.592199] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1040.592760] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1040.598134] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1040.598134] env[62600]: value = "task-1222984" [ 1040.598134] env[62600]: _type = "Task" [ 1040.598134] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.606861] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222984, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.626982] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "1817183b-44a5-41ec-a2bf-818edee264b3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.627293] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "1817183b-44a5-41ec-a2bf-818edee264b3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.720174] env[62600]: DEBUG nova.compute.manager [req-265a1886-881d-4a42-ab9a-c434546478ab req-c3313b63-ceaa-4858-99c5-44c6d376a7ca service nova] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Received event network-changed-acb612e6-f57e-4eae-8764-c9525e4c6baa {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1040.720174] env[62600]: DEBUG nova.compute.manager [req-265a1886-881d-4a42-ab9a-c434546478ab req-c3313b63-ceaa-4858-99c5-44c6d376a7ca service nova] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Refreshing instance network info cache due to event network-changed-acb612e6-f57e-4eae-8764-c9525e4c6baa. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1040.720174] env[62600]: DEBUG oslo_concurrency.lockutils [req-265a1886-881d-4a42-ab9a-c434546478ab req-c3313b63-ceaa-4858-99c5-44c6d376a7ca service nova] Acquiring lock "refresh_cache-4607dacd-28ab-4173-ba86-8a78bec19dbc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.720369] env[62600]: DEBUG oslo_concurrency.lockutils [req-265a1886-881d-4a42-ab9a-c434546478ab req-c3313b63-ceaa-4858-99c5-44c6d376a7ca service nova] Acquired lock "refresh_cache-4607dacd-28ab-4173-ba86-8a78bec19dbc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.720598] env[62600]: DEBUG nova.network.neutron [req-265a1886-881d-4a42-ab9a-c434546478ab req-c3313b63-ceaa-4858-99c5-44c6d376a7ca service nova] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Refreshing network info cache for port acb612e6-f57e-4eae-8764-c9525e4c6baa {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1040.752988] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba2c114-4022-4a80-b06e-afd756f963cb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.763596] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-698c8c03-37fd-4c00-b4e8-5ed9839776a7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.793489] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144a4ad4-f7dc-49f5-ad6a-fce39a00d390 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.801746] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a6f9fa-ced8-4d71-b209-2a64ce7cb945 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.815146] env[62600]: DEBUG nova.compute.provider_tree [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1040.869113] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222983, 'name': ReconfigVM_Task, 'duration_secs': 0.345162} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.869405] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 8417575c-d68d-464f-8c22-ab358deb76d8/8417575c-d68d-464f-8c22-ab358deb76d8.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1040.870132] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fd814008-06df-4082-a5c1-8ccca91bd577 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.876886] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1040.876886] env[62600]: value = "task-1222985" [ 1040.876886] env[62600]: _type = "Task" [ 1040.876886] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.884863] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222985, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.108340] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222984, 'name': CreateVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.129773] env[62600]: DEBUG nova.compute.manager [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1041.318438] env[62600]: DEBUG nova.scheduler.client.report [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1041.387938] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222985, 'name': Rename_Task, 'duration_secs': 0.14097} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.388297] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1041.388582] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-80030991-c3f5-4e13-8384-1921aa465aa1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.394723] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1041.394723] env[62600]: value = "task-1222986" [ 1041.394723] env[62600]: _type = "Task" [ 1041.394723] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.402946] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222986, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.483203] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "0416ef03-fd45-405f-9047-76b34a839217" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.483588] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "0416ef03-fd45-405f-9047-76b34a839217" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.483839] env[62600]: DEBUG nova.compute.manager [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Going to confirm migration 4 {{(pid=62600) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1041.495115] env[62600]: DEBUG nova.network.neutron [req-265a1886-881d-4a42-ab9a-c434546478ab req-c3313b63-ceaa-4858-99c5-44c6d376a7ca service nova] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Updated VIF entry in instance network info cache for port acb612e6-f57e-4eae-8764-c9525e4c6baa. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1041.495115] env[62600]: DEBUG nova.network.neutron [req-265a1886-881d-4a42-ab9a-c434546478ab req-c3313b63-ceaa-4858-99c5-44c6d376a7ca service nova] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Updating instance_info_cache with network_info: [{"id": "acb612e6-f57e-4eae-8764-c9525e4c6baa", "address": "fa:16:3e:ef:40:fe", "network": {"id": "1d2052cf-5be1-48e7-887a-18f367b7823f", "bridge": "br-int", "label": "tempest-ServersTestJSON-106099679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f91091f83ee4a2091507ca994e3d52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapacb612e6-f5", "ovs_interfaceid": "acb612e6-f57e-4eae-8764-c9525e4c6baa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.608872] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222984, 'name': CreateVM_Task, 'duration_secs': 0.56408} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.609044] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1041.609731] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.609902] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.610244] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1041.610497] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a007618c-5e7b-4999-a2c0-a88ee8eabd73 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.614890] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1041.614890] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52850da2-9825-5ae9-f4aa-58a959db7ad3" [ 1041.614890] env[62600]: _type = "Task" [ 1041.614890] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.622248] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52850da2-9825-5ae9-f4aa-58a959db7ad3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.651883] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.727744] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1041.728036] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1041.728233] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Starting heal instance info cache {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1041.823886] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.258s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.824555] env[62600]: DEBUG nova.compute.manager [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1041.827405] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.011s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.827656] env[62600]: DEBUG nova.objects.instance [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lazy-loading 'resources' on Instance uuid d66c404d-8fd1-4fb7-a3b9-f21854c7e735 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.907324] env[62600]: DEBUG oslo_vmware.api [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1222986, 'name': PowerOnVM_Task, 'duration_secs': 0.497566} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.907324] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1041.907324] env[62600]: INFO nova.compute.manager [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Took 6.75 seconds to spawn the instance on the hypervisor. [ 1041.907324] env[62600]: DEBUG nova.compute.manager [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1041.907324] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-083ce75b-5f57-4f74-b3d6-fbb9a045856f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.996339] env[62600]: DEBUG oslo_concurrency.lockutils [req-265a1886-881d-4a42-ab9a-c434546478ab req-c3313b63-ceaa-4858-99c5-44c6d376a7ca service nova] Releasing lock "refresh_cache-4607dacd-28ab-4173-ba86-8a78bec19dbc" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.996648] env[62600]: DEBUG nova.compute.manager [req-265a1886-881d-4a42-ab9a-c434546478ab req-c3313b63-ceaa-4858-99c5-44c6d376a7ca service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Received event network-changed-b8e9cd2e-1990-4997-b1fa-cb4411d39fde {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1041.996836] env[62600]: DEBUG nova.compute.manager [req-265a1886-881d-4a42-ab9a-c434546478ab req-c3313b63-ceaa-4858-99c5-44c6d376a7ca service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Refreshing instance network info cache due to event network-changed-b8e9cd2e-1990-4997-b1fa-cb4411d39fde. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1041.997059] env[62600]: DEBUG oslo_concurrency.lockutils [req-265a1886-881d-4a42-ab9a-c434546478ab req-c3313b63-ceaa-4858-99c5-44c6d376a7ca service nova] Acquiring lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.997216] env[62600]: DEBUG oslo_concurrency.lockutils [req-265a1886-881d-4a42-ab9a-c434546478ab req-c3313b63-ceaa-4858-99c5-44c6d376a7ca service nova] Acquired lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.997383] env[62600]: DEBUG nova.network.neutron [req-265a1886-881d-4a42-ab9a-c434546478ab req-c3313b63-ceaa-4858-99c5-44c6d376a7ca service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Refreshing network info cache for port b8e9cd2e-1990-4997-b1fa-cb4411d39fde {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1042.040721] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "refresh_cache-0416ef03-fd45-405f-9047-76b34a839217" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.040917] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "refresh_cache-0416ef03-fd45-405f-9047-76b34a839217" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.041110] env[62600]: DEBUG nova.network.neutron [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1042.041313] env[62600]: DEBUG nova.objects.instance [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lazy-loading 'info_cache' on Instance uuid 0416ef03-fd45-405f-9047-76b34a839217 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.125857] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52850da2-9825-5ae9-f4aa-58a959db7ad3, 'name': SearchDatastore_Task, 'duration_secs': 0.01844} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.126314] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.126633] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1042.126954] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.127226] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.127501] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1042.127849] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-27f6aa22-dd66-428c-b1ad-393be0e47139 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.136132] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1042.136401] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1042.137222] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e653b74e-bba0-4666-9c66-65b6d218b613 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.142511] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1042.142511] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52eee0c2-0d7e-7fb4-21b4-6dde00f63c2a" [ 1042.142511] env[62600]: _type = "Task" [ 1042.142511] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.151770] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52eee0c2-0d7e-7fb4-21b4-6dde00f63c2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.230970] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Didn't find any instances for network info cache update. {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1042.231246] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1042.231389] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62600) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1042.330838] env[62600]: DEBUG nova.compute.utils [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1042.334895] env[62600]: DEBUG nova.compute.manager [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1042.335067] env[62600]: DEBUG nova.network.neutron [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1042.372978] env[62600]: DEBUG nova.policy [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '57644cb1856e4efba88d25065ce1dc68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '952323d62f034126a9b46750d03fe6ea', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 1042.425474] env[62600]: INFO nova.compute.manager [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Took 15.99 seconds to build instance. [ 1042.465176] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-042c2a17-6db7-41d9-9df3-8799bbd34925 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.472570] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f98ce3e-9e62-48b8-b80d-22afb694ef38 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.505356] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d0c905-0b1b-4d79-ab67-cb08a9b97f39 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.513170] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c87e85f-619c-4c9d-88ca-15863ed91a1b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.526321] env[62600]: DEBUG nova.compute.provider_tree [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.653698] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52eee0c2-0d7e-7fb4-21b4-6dde00f63c2a, 'name': SearchDatastore_Task, 'duration_secs': 0.008098} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.654548] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b38d4c2-9bfe-48e3-93d0-11294f8b0c32 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.660250] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1042.660250] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526d4017-2693-6f81-45e8-65df5545f824" [ 1042.660250] env[62600]: _type = "Task" [ 1042.660250] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.672159] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526d4017-2693-6f81-45e8-65df5545f824, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.743826] env[62600]: DEBUG nova.network.neutron [req-265a1886-881d-4a42-ab9a-c434546478ab req-c3313b63-ceaa-4858-99c5-44c6d376a7ca service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Updated VIF entry in instance network info cache for port b8e9cd2e-1990-4997-b1fa-cb4411d39fde. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1042.744181] env[62600]: DEBUG nova.network.neutron [req-265a1886-881d-4a42-ab9a-c434546478ab req-c3313b63-ceaa-4858-99c5-44c6d376a7ca service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Updating instance_info_cache with network_info: [{"id": "b8e9cd2e-1990-4997-b1fa-cb4411d39fde", "address": "fa:16:3e:fc:c1:ac", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8e9cd2e-19", "ovs_interfaceid": "b8e9cd2e-1990-4997-b1fa-cb4411d39fde", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.828705] env[62600]: DEBUG nova.network.neutron [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Successfully created port: 280baabd-3083-4f33-b294-1f988193665b {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1042.836513] env[62600]: DEBUG nova.compute.manager [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1042.927815] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fb2d60e3-fa60-4915-865d-107349dfd91e tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "8417575c-d68d-464f-8c22-ab358deb76d8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.504s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.030051] env[62600]: DEBUG nova.scheduler.client.report [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1043.171874] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526d4017-2693-6f81-45e8-65df5545f824, 'name': SearchDatastore_Task, 'duration_secs': 0.010786} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.172275] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.172572] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 4607dacd-28ab-4173-ba86-8a78bec19dbc/4607dacd-28ab-4173-ba86-8a78bec19dbc.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1043.172844] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4f43e3dc-ce4c-4949-a00d-c1b359dbfb1d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.179142] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1043.179142] env[62600]: value = "task-1222987" [ 1043.179142] env[62600]: _type = "Task" [ 1043.179142] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.186782] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222987, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.239570] env[62600]: DEBUG nova.network.neutron [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Updating instance_info_cache with network_info: [{"id": "bb15b28d-be8f-4fff-b8a6-0801d3e0541a", "address": "fa:16:3e:b1:43:47", "network": {"id": "95818650-8227-4d11-9d14-007f06c07790", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1738152115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73e664d178f7484a9f4741b4d9450e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb15b28d-be", "ovs_interfaceid": "bb15b28d-be8f-4fff-b8a6-0801d3e0541a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.248786] env[62600]: DEBUG oslo_concurrency.lockutils [req-265a1886-881d-4a42-ab9a-c434546478ab req-c3313b63-ceaa-4858-99c5-44c6d376a7ca service nova] Releasing lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.535452] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.708s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.538789] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.886s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.539869] env[62600]: INFO nova.compute.claims [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1043.561287] env[62600]: INFO nova.scheduler.client.report [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Deleted allocations for instance d66c404d-8fd1-4fb7-a3b9-f21854c7e735 [ 1043.598799] env[62600]: DEBUG nova.compute.manager [req-25b42487-a6ae-4205-abef-6307bab7cedd req-bdb9f2e5-b219-461e-9792-c8f4f0c306a2 service nova] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Received event network-changed-028950a6-9715-4e67-b31b-447158a62875 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1043.599094] env[62600]: DEBUG nova.compute.manager [req-25b42487-a6ae-4205-abef-6307bab7cedd req-bdb9f2e5-b219-461e-9792-c8f4f0c306a2 service nova] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Refreshing instance network info cache due to event network-changed-028950a6-9715-4e67-b31b-447158a62875. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1043.599333] env[62600]: DEBUG oslo_concurrency.lockutils [req-25b42487-a6ae-4205-abef-6307bab7cedd req-bdb9f2e5-b219-461e-9792-c8f4f0c306a2 service nova] Acquiring lock "refresh_cache-8417575c-d68d-464f-8c22-ab358deb76d8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.599448] env[62600]: DEBUG oslo_concurrency.lockutils [req-25b42487-a6ae-4205-abef-6307bab7cedd req-bdb9f2e5-b219-461e-9792-c8f4f0c306a2 service nova] Acquired lock "refresh_cache-8417575c-d68d-464f-8c22-ab358deb76d8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.599662] env[62600]: DEBUG nova.network.neutron [req-25b42487-a6ae-4205-abef-6307bab7cedd req-bdb9f2e5-b219-461e-9792-c8f4f0c306a2 service nova] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Refreshing network info cache for port 028950a6-9715-4e67-b31b-447158a62875 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1043.689462] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222987, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.436232} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.689770] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 4607dacd-28ab-4173-ba86-8a78bec19dbc/4607dacd-28ab-4173-ba86-8a78bec19dbc.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1043.689996] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1043.690271] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8c0ad434-7e19-4a0a-9bb3-b1617d3d3e42 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.696888] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1043.696888] env[62600]: value = "task-1222988" [ 1043.696888] env[62600]: _type = "Task" [ 1043.696888] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.704466] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222988, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.742827] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "refresh_cache-0416ef03-fd45-405f-9047-76b34a839217" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.743212] env[62600]: DEBUG nova.objects.instance [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lazy-loading 'migration_context' on Instance uuid 0416ef03-fd45-405f-9047-76b34a839217 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.845875] env[62600]: DEBUG nova.compute.manager [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1043.870452] env[62600]: DEBUG nova.virt.hardware [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1043.870709] env[62600]: DEBUG nova.virt.hardware [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1043.870873] env[62600]: DEBUG nova.virt.hardware [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1043.871073] env[62600]: DEBUG nova.virt.hardware [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1043.871299] env[62600]: DEBUG nova.virt.hardware [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1043.871467] env[62600]: DEBUG nova.virt.hardware [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1043.871680] env[62600]: DEBUG nova.virt.hardware [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1043.871849] env[62600]: DEBUG nova.virt.hardware [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1043.872033] env[62600]: DEBUG nova.virt.hardware [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1043.872207] env[62600]: DEBUG nova.virt.hardware [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1043.872388] env[62600]: DEBUG nova.virt.hardware [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1043.873268] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-733b2c9c-e100-4d38-b273-1d72e996b0f8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.881097] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0014ef3-8e5e-44ea-9f67-d3575b71c06f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.072538] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a85d6631-b5d9-40f7-abd1-0a9ca4416cf4 tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "d66c404d-8fd1-4fb7-a3b9-f21854c7e735" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.194s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.208629] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222988, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056739} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.208962] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1044.209729] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b84188e5-e919-49ce-ad57-c5b45d6b415a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.231206] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 4607dacd-28ab-4173-ba86-8a78bec19dbc/4607dacd-28ab-4173-ba86-8a78bec19dbc.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1044.233465] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5944ad9-6e73-49cd-8b57-80e447570c9c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.247890] env[62600]: DEBUG nova.objects.base [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Object Instance<0416ef03-fd45-405f-9047-76b34a839217> lazy-loaded attributes: info_cache,migration_context {{(pid=62600) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1044.248890] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9b0b3d-2b37-40f1-b41f-48083ac3c2c3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.268437] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d0ed7a2-2b9f-4037-a802-938d46a45906 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.270758] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1044.270758] env[62600]: value = "task-1222989" [ 1044.270758] env[62600]: _type = "Task" [ 1044.270758] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.274878] env[62600]: DEBUG oslo_vmware.api [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1044.274878] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e2d189-9535-f4f4-c998-730edf96702c" [ 1044.274878] env[62600]: _type = "Task" [ 1044.274878] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.284129] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222989, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.288687] env[62600]: DEBUG oslo_vmware.api [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e2d189-9535-f4f4-c998-730edf96702c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.346888] env[62600]: DEBUG nova.network.neutron [req-25b42487-a6ae-4205-abef-6307bab7cedd req-bdb9f2e5-b219-461e-9792-c8f4f0c306a2 service nova] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Updated VIF entry in instance network info cache for port 028950a6-9715-4e67-b31b-447158a62875. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1044.347296] env[62600]: DEBUG nova.network.neutron [req-25b42487-a6ae-4205-abef-6307bab7cedd req-bdb9f2e5-b219-461e-9792-c8f4f0c306a2 service nova] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Updating instance_info_cache with network_info: [{"id": "028950a6-9715-4e67-b31b-447158a62875", "address": "fa:16:3e:f9:ef:88", "network": {"id": "e7547ac4-24a6-4778-a17a-40e79e7a0d0c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-320671681-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.155", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3bb131f3994f7a8673dafff6d353c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap028950a6-97", "ovs_interfaceid": "028950a6-9715-4e67-b31b-447158a62875", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.688695] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57702034-7f9a-4ba0-b0f4-46ea70bc3e7c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.696212] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6b13951-b1e4-499c-8423-525e338423bd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.728755] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc12b506-06ec-4f6f-b148-ffc1c09a106d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.736848] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e0cbc50-caca-465e-85ec-e576b81d400f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.751594] env[62600]: DEBUG nova.compute.provider_tree [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1044.781405] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222989, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.789830] env[62600]: DEBUG oslo_vmware.api [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e2d189-9535-f4f4-c998-730edf96702c, 'name': SearchDatastore_Task, 'duration_secs': 0.018328} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.790207] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.850381] env[62600]: DEBUG oslo_concurrency.lockutils [req-25b42487-a6ae-4205-abef-6307bab7cedd req-bdb9f2e5-b219-461e-9792-c8f4f0c306a2 service nova] Releasing lock "refresh_cache-8417575c-d68d-464f-8c22-ab358deb76d8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.851625] env[62600]: DEBUG nova.network.neutron [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Successfully updated port: 280baabd-3083-4f33-b294-1f988193665b {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1045.026515] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "b967fb11-e70e-4e17-b769-38da581bd83b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.026748] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "b967fb11-e70e-4e17-b769-38da581bd83b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.026964] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "b967fb11-e70e-4e17-b769-38da581bd83b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.027180] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "b967fb11-e70e-4e17-b769-38da581bd83b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.027364] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "b967fb11-e70e-4e17-b769-38da581bd83b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.029482] env[62600]: INFO nova.compute.manager [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Terminating instance [ 1045.032026] env[62600]: DEBUG nova.compute.manager [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1045.032026] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1045.032280] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd109b84-388d-42e6-929f-f45177ebcbf8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.039748] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1045.039983] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f24f5549-9721-4399-bbd3-e620d1d28caa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.047150] env[62600]: DEBUG oslo_vmware.api [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 1045.047150] env[62600]: value = "task-1222990" [ 1045.047150] env[62600]: _type = "Task" [ 1045.047150] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.054299] env[62600]: DEBUG oslo_vmware.api [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222990, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.255678] env[62600]: DEBUG nova.scheduler.client.report [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1045.282187] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222989, 'name': ReconfigVM_Task, 'duration_secs': 0.606791} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.282416] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 4607dacd-28ab-4173-ba86-8a78bec19dbc/4607dacd-28ab-4173-ba86-8a78bec19dbc.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1045.283055] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57c02e37-24ec-490b-b8b2-a4452321fa73 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.288847] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1045.288847] env[62600]: value = "task-1222991" [ 1045.288847] env[62600]: _type = "Task" [ 1045.288847] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.297240] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222991, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.357973] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.357973] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquired lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.357973] env[62600]: DEBUG nova.network.neutron [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1045.559828] env[62600]: DEBUG oslo_vmware.api [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222990, 'name': PowerOffVM_Task, 'duration_secs': 0.274041} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.559828] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1045.559828] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1045.559828] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-12911f4b-0122-49dd-b60f-2a090e37a065 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.632546] env[62600]: DEBUG nova.compute.manager [req-e2ddcf29-6344-41e6-9191-d34eff9c9a2c req-8ab75935-28a4-455d-8570-37f3c58fe7cb service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Received event network-vif-plugged-280baabd-3083-4f33-b294-1f988193665b {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1045.632875] env[62600]: DEBUG oslo_concurrency.lockutils [req-e2ddcf29-6344-41e6-9191-d34eff9c9a2c req-8ab75935-28a4-455d-8570-37f3c58fe7cb service nova] Acquiring lock "692fa608-39f4-4d7d-9f30-b6559878b970-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.633292] env[62600]: DEBUG oslo_concurrency.lockutils [req-e2ddcf29-6344-41e6-9191-d34eff9c9a2c req-8ab75935-28a4-455d-8570-37f3c58fe7cb service nova] Lock "692fa608-39f4-4d7d-9f30-b6559878b970-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.633470] env[62600]: DEBUG oslo_concurrency.lockutils [req-e2ddcf29-6344-41e6-9191-d34eff9c9a2c req-8ab75935-28a4-455d-8570-37f3c58fe7cb service nova] Lock "692fa608-39f4-4d7d-9f30-b6559878b970-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.633709] env[62600]: DEBUG nova.compute.manager [req-e2ddcf29-6344-41e6-9191-d34eff9c9a2c req-8ab75935-28a4-455d-8570-37f3c58fe7cb service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] No waiting events found dispatching network-vif-plugged-280baabd-3083-4f33-b294-1f988193665b {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1045.633944] env[62600]: WARNING nova.compute.manager [req-e2ddcf29-6344-41e6-9191-d34eff9c9a2c req-8ab75935-28a4-455d-8570-37f3c58fe7cb service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Received unexpected event network-vif-plugged-280baabd-3083-4f33-b294-1f988193665b for instance with vm_state building and task_state spawning. [ 1045.634217] env[62600]: DEBUG nova.compute.manager [req-e2ddcf29-6344-41e6-9191-d34eff9c9a2c req-8ab75935-28a4-455d-8570-37f3c58fe7cb service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Received event network-changed-280baabd-3083-4f33-b294-1f988193665b {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1045.634452] env[62600]: DEBUG nova.compute.manager [req-e2ddcf29-6344-41e6-9191-d34eff9c9a2c req-8ab75935-28a4-455d-8570-37f3c58fe7cb service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Refreshing instance network info cache due to event network-changed-280baabd-3083-4f33-b294-1f988193665b. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1045.634691] env[62600]: DEBUG oslo_concurrency.lockutils [req-e2ddcf29-6344-41e6-9191-d34eff9c9a2c req-8ab75935-28a4-455d-8570-37f3c58fe7cb service nova] Acquiring lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.636997] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1045.637294] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1045.637553] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Deleting the datastore file [datastore2] b967fb11-e70e-4e17-b769-38da581bd83b {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1045.637876] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-57a3a777-46a1-4b34-b2b8-d40e6aae5aa5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.646087] env[62600]: DEBUG oslo_vmware.api [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for the task: (returnval){ [ 1045.646087] env[62600]: value = "task-1222993" [ 1045.646087] env[62600]: _type = "Task" [ 1045.646087] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.656755] env[62600]: DEBUG oslo_vmware.api [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222993, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.761307] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.223s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.761922] env[62600]: DEBUG nova.compute.manager [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1045.766236] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.976s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.799063] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222991, 'name': Rename_Task, 'duration_secs': 0.137069} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.799336] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1045.799594] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c1c3376d-9956-46a4-a456-64fc8e251a84 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.806227] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1045.806227] env[62600]: value = "task-1222994" [ 1045.806227] env[62600]: _type = "Task" [ 1045.806227] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.813987] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222994, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.931830] env[62600]: DEBUG nova.network.neutron [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1046.142578] env[62600]: DEBUG nova.network.neutron [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Updating instance_info_cache with network_info: [{"id": "280baabd-3083-4f33-b294-1f988193665b", "address": "fa:16:3e:72:db:a2", "network": {"id": "dc21424f-b891-428d-a74e-76ab78d757a8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1428037886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "952323d62f034126a9b46750d03fe6ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap280baabd-30", "ovs_interfaceid": "280baabd-3083-4f33-b294-1f988193665b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.155630] env[62600]: DEBUG oslo_vmware.api [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Task: {'id': task-1222993, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.319676} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.156733] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1046.156932] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1046.157132] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1046.157312] env[62600]: INFO nova.compute.manager [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1046.157593] env[62600]: DEBUG oslo.service.loopingcall [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1046.157978] env[62600]: DEBUG nova.compute.manager [-] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1046.158087] env[62600]: DEBUG nova.network.neutron [-] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1046.276039] env[62600]: DEBUG nova.compute.utils [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1046.276039] env[62600]: DEBUG nova.compute.manager [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1046.276039] env[62600]: DEBUG nova.network.neutron [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1046.316617] env[62600]: DEBUG nova.policy [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fbd1b2fb34d841359ada8fc44bec2986', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b028450f2da445fb83e37adfc86bba68', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 1046.323420] env[62600]: DEBUG oslo_vmware.api [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222994, 'name': PowerOnVM_Task, 'duration_secs': 0.47012} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.323916] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1046.324184] env[62600]: INFO nova.compute.manager [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Took 8.00 seconds to spawn the instance on the hypervisor. [ 1046.324573] env[62600]: DEBUG nova.compute.manager [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1046.325307] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d395a7-91d9-4891-a830-98d2c06122c1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.433050] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-407138da-563b-4768-bd3d-b42586408a8f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.440576] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-989d2efd-5511-40dc-91e9-77e126ed1779 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.472127] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc4c428c-4f6b-4148-b97b-029cca6341a9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.479501] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac8f8fa8-a83d-4c24-b6ed-ec085f75c9b5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.492844] env[62600]: DEBUG nova.compute.provider_tree [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.587277] env[62600]: DEBUG nova.network.neutron [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Successfully created port: 5796bb3c-0926-47b4-bcdb-c54c18f83200 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1046.645567] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Releasing lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.645897] env[62600]: DEBUG nova.compute.manager [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Instance network_info: |[{"id": "280baabd-3083-4f33-b294-1f988193665b", "address": "fa:16:3e:72:db:a2", "network": {"id": "dc21424f-b891-428d-a74e-76ab78d757a8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1428037886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "952323d62f034126a9b46750d03fe6ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap280baabd-30", "ovs_interfaceid": "280baabd-3083-4f33-b294-1f988193665b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1046.646254] env[62600]: DEBUG oslo_concurrency.lockutils [req-e2ddcf29-6344-41e6-9191-d34eff9c9a2c req-8ab75935-28a4-455d-8570-37f3c58fe7cb service nova] Acquired lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.646440] env[62600]: DEBUG nova.network.neutron [req-e2ddcf29-6344-41e6-9191-d34eff9c9a2c req-8ab75935-28a4-455d-8570-37f3c58fe7cb service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Refreshing network info cache for port 280baabd-3083-4f33-b294-1f988193665b {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1046.647615] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:72:db:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '07e9bef1-2b0e-4e4d-997f-de71bb0e213a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '280baabd-3083-4f33-b294-1f988193665b', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1046.654972] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Creating folder: Project (952323d62f034126a9b46750d03fe6ea). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1046.658402] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a31ff1c9-0106-438e-8974-586f7175ae1e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.669792] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Created folder: Project (952323d62f034126a9b46750d03fe6ea) in parent group-v264198. [ 1046.670857] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Creating folder: Instances. Parent ref: group-v264366. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1046.670857] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-af007c82-eb12-4e24-935f-524abb75cb6c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.679098] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Created folder: Instances in parent group-v264366. [ 1046.679358] env[62600]: DEBUG oslo.service.loopingcall [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1046.679556] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1046.679918] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a5c664b7-2da0-47bd-9346-5571bf321804 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.699926] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1046.699926] env[62600]: value = "task-1222997" [ 1046.699926] env[62600]: _type = "Task" [ 1046.699926] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.706986] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222997, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.778981] env[62600]: DEBUG nova.compute.manager [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1046.845913] env[62600]: INFO nova.compute.manager [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Took 12.84 seconds to build instance. [ 1046.882448] env[62600]: DEBUG nova.network.neutron [-] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.937872] env[62600]: DEBUG nova.network.neutron [req-e2ddcf29-6344-41e6-9191-d34eff9c9a2c req-8ab75935-28a4-455d-8570-37f3c58fe7cb service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Updated VIF entry in instance network info cache for port 280baabd-3083-4f33-b294-1f988193665b. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1046.938275] env[62600]: DEBUG nova.network.neutron [req-e2ddcf29-6344-41e6-9191-d34eff9c9a2c req-8ab75935-28a4-455d-8570-37f3c58fe7cb service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Updating instance_info_cache with network_info: [{"id": "280baabd-3083-4f33-b294-1f988193665b", "address": "fa:16:3e:72:db:a2", "network": {"id": "dc21424f-b891-428d-a74e-76ab78d757a8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1428037886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "952323d62f034126a9b46750d03fe6ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap280baabd-30", "ovs_interfaceid": "280baabd-3083-4f33-b294-1f988193665b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.995735] env[62600]: DEBUG nova.scheduler.client.report [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1047.209520] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1222997, 'name': CreateVM_Task, 'duration_secs': 0.306886} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.209712] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1047.210372] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.210548] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.210878] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1047.211149] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a0bf2d3-c282-42e0-8a9e-abadcd3792b4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.215521] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1047.215521] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52808bd5-0dc5-7219-5eea-b363403db79d" [ 1047.215521] env[62600]: _type = "Task" [ 1047.215521] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.224258] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52808bd5-0dc5-7219-5eea-b363403db79d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.348393] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2658489b-52e3-43e4-a6af-99f6f6b49955 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "4607dacd-28ab-4173-ba86-8a78bec19dbc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.347s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.388286] env[62600]: INFO nova.compute.manager [-] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Took 1.23 seconds to deallocate network for instance. [ 1047.440808] env[62600]: DEBUG oslo_concurrency.lockutils [req-e2ddcf29-6344-41e6-9191-d34eff9c9a2c req-8ab75935-28a4-455d-8570-37f3c58fe7cb service nova] Releasing lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.655706] env[62600]: DEBUG nova.compute.manager [req-6f3f526c-1480-43b2-af6c-44d63945de9c req-bbd36378-0a9d-453a-a46b-96dd30bec8e1 service nova] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Received event network-vif-deleted-d6e6a2fb-7b6b-42bd-8b1a-c8026b45d049 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1047.725580] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52808bd5-0dc5-7219-5eea-b363403db79d, 'name': SearchDatastore_Task, 'duration_secs': 0.009139} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.725891] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.726158] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1047.726427] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.726590] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.726781] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1047.727040] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-73e42883-272a-4a47-bf92-bd2c86e9b6b8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.734915] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1047.735107] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1047.735829] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9d4b276-279d-43c4-bd24-24cd4d5ada1f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.740920] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1047.740920] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5207a21f-52bd-af46-601a-416157175ef4" [ 1047.740920] env[62600]: _type = "Task" [ 1047.740920] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.747853] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5207a21f-52bd-af46-601a-416157175ef4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.790638] env[62600]: DEBUG nova.compute.manager [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1047.814580] env[62600]: DEBUG nova.virt.hardware [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1047.814868] env[62600]: DEBUG nova.virt.hardware [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1047.814981] env[62600]: DEBUG nova.virt.hardware [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1047.815233] env[62600]: DEBUG nova.virt.hardware [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1047.815374] env[62600]: DEBUG nova.virt.hardware [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1047.815532] env[62600]: DEBUG nova.virt.hardware [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1047.815809] env[62600]: DEBUG nova.virt.hardware [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1047.815906] env[62600]: DEBUG nova.virt.hardware [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1047.816091] env[62600]: DEBUG nova.virt.hardware [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1047.816262] env[62600]: DEBUG nova.virt.hardware [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1047.816440] env[62600]: DEBUG nova.virt.hardware [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1047.817571] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbfae4b8-fe11-4528-ba11-6be4bf0e52a2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.825313] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f1463f-9f5d-4d05-8314-7170595100d7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.894295] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.970965] env[62600]: DEBUG nova.compute.manager [req-82f746e2-3ce8-4371-8950-da66eb9aef65 req-3550c593-9184-4be1-93f3-0d80f220a1ce service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Received event network-vif-plugged-5796bb3c-0926-47b4-bcdb-c54c18f83200 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1047.971484] env[62600]: DEBUG oslo_concurrency.lockutils [req-82f746e2-3ce8-4371-8950-da66eb9aef65 req-3550c593-9184-4be1-93f3-0d80f220a1ce service nova] Acquiring lock "1817183b-44a5-41ec-a2bf-818edee264b3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.971843] env[62600]: DEBUG oslo_concurrency.lockutils [req-82f746e2-3ce8-4371-8950-da66eb9aef65 req-3550c593-9184-4be1-93f3-0d80f220a1ce service nova] Lock "1817183b-44a5-41ec-a2bf-818edee264b3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.972159] env[62600]: DEBUG oslo_concurrency.lockutils [req-82f746e2-3ce8-4371-8950-da66eb9aef65 req-3550c593-9184-4be1-93f3-0d80f220a1ce service nova] Lock "1817183b-44a5-41ec-a2bf-818edee264b3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.972482] env[62600]: DEBUG nova.compute.manager [req-82f746e2-3ce8-4371-8950-da66eb9aef65 req-3550c593-9184-4be1-93f3-0d80f220a1ce service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] No waiting events found dispatching network-vif-plugged-5796bb3c-0926-47b4-bcdb-c54c18f83200 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1047.972785] env[62600]: WARNING nova.compute.manager [req-82f746e2-3ce8-4371-8950-da66eb9aef65 req-3550c593-9184-4be1-93f3-0d80f220a1ce service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Received unexpected event network-vif-plugged-5796bb3c-0926-47b4-bcdb-c54c18f83200 for instance with vm_state building and task_state spawning. [ 1048.007536] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.241s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.011424] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.117s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.013593] env[62600]: DEBUG nova.objects.instance [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lazy-loading 'resources' on Instance uuid b967fb11-e70e-4e17-b769-38da581bd83b {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1048.050782] env[62600]: DEBUG oslo_concurrency.lockutils [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "4607dacd-28ab-4173-ba86-8a78bec19dbc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.050782] env[62600]: DEBUG oslo_concurrency.lockutils [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "4607dacd-28ab-4173-ba86-8a78bec19dbc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.050782] env[62600]: DEBUG oslo_concurrency.lockutils [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "4607dacd-28ab-4173-ba86-8a78bec19dbc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.050782] env[62600]: DEBUG oslo_concurrency.lockutils [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "4607dacd-28ab-4173-ba86-8a78bec19dbc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.050782] env[62600]: DEBUG oslo_concurrency.lockutils [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "4607dacd-28ab-4173-ba86-8a78bec19dbc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.051337] env[62600]: INFO nova.compute.manager [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Terminating instance [ 1048.053318] env[62600]: DEBUG nova.compute.manager [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1048.053354] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1048.057174] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-161501b4-f7b2-4c92-8208-e19b4324b59d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.062142] env[62600]: DEBUG nova.network.neutron [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Successfully updated port: 5796bb3c-0926-47b4-bcdb-c54c18f83200 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1048.068900] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1048.069184] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01236fd5-cc40-4234-896c-940aacb6a324 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.077233] env[62600]: DEBUG oslo_vmware.api [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1048.077233] env[62600]: value = "task-1222998" [ 1048.077233] env[62600]: _type = "Task" [ 1048.077233] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.086020] env[62600]: DEBUG oslo_vmware.api [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222998, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.254152] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5207a21f-52bd-af46-601a-416157175ef4, 'name': SearchDatastore_Task, 'duration_secs': 0.008258} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.255071] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c911d3a8-dee3-40d4-b9bf-3cc70d804824 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.260184] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1048.260184] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5287b730-ff27-caef-a1ff-c7158fd4aa63" [ 1048.260184] env[62600]: _type = "Task" [ 1048.260184] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.267577] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5287b730-ff27-caef-a1ff-c7158fd4aa63, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.565555] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.565887] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.565887] env[62600]: DEBUG nova.network.neutron [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1048.584380] env[62600]: INFO nova.scheduler.client.report [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleted allocation for migration a4591da5-9da8-4860-abec-9a3b0513c9a2 [ 1048.588427] env[62600]: DEBUG oslo_vmware.api [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1222998, 'name': PowerOffVM_Task, 'duration_secs': 0.200014} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.593048] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1048.593048] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1048.593620] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7cc37552-6a01-4f58-a678-d80a12add6aa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.653633] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1048.653872] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1048.654075] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleting the datastore file [datastore1] 4607dacd-28ab-4173-ba86-8a78bec19dbc {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1048.654344] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1339ba12-47bf-435d-8205-1e69593eb0b3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.658223] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7855117-cab0-4211-be60-01744c35fca3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.661527] env[62600]: DEBUG oslo_vmware.api [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1048.661527] env[62600]: value = "task-1223000" [ 1048.661527] env[62600]: _type = "Task" [ 1048.661527] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.668925] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39cb6f27-1451-43fb-aae2-6f70779e12bf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.674555] env[62600]: DEBUG oslo_vmware.api [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223000, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.702265] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0706ad3a-4c10-4701-8bd8-19a04fba0d36 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.709155] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be7ea8ab-696d-4ba7-9e2e-480f9e5fd899 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.723155] env[62600]: DEBUG nova.compute.provider_tree [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1048.770346] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5287b730-ff27-caef-a1ff-c7158fd4aa63, 'name': SearchDatastore_Task, 'duration_secs': 0.008603} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.770623] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.770874] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 692fa608-39f4-4d7d-9f30-b6559878b970/692fa608-39f4-4d7d-9f30-b6559878b970.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1048.771357] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-69615e04-40f5-473e-b32e-706d94424405 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.777862] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1048.777862] env[62600]: value = "task-1223001" [ 1048.777862] env[62600]: _type = "Task" [ 1048.777862] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.785111] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223001, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.095345] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5888cdb1-1431-4d3b-9cf8-353e2e906609 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "0416ef03-fd45-405f-9047-76b34a839217" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.612s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.116638] env[62600]: DEBUG nova.network.neutron [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1049.175627] env[62600]: DEBUG oslo_vmware.api [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223000, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145457} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.176333] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1049.176333] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1049.176550] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1049.176798] env[62600]: INFO nova.compute.manager [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1049.177207] env[62600]: DEBUG oslo.service.loopingcall [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1049.177527] env[62600]: DEBUG nova.compute.manager [-] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1049.177684] env[62600]: DEBUG nova.network.neutron [-] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1049.226532] env[62600]: DEBUG nova.scheduler.client.report [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1049.268628] env[62600]: DEBUG nova.network.neutron [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Updating instance_info_cache with network_info: [{"id": "5796bb3c-0926-47b4-bcdb-c54c18f83200", "address": "fa:16:3e:94:99:6b", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5796bb3c-09", "ovs_interfaceid": "5796bb3c-0926-47b4-bcdb-c54c18f83200", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.287431] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223001, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.438559} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.287782] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 692fa608-39f4-4d7d-9f30-b6559878b970/692fa608-39f4-4d7d-9f30-b6559878b970.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1049.287999] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1049.288266] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-137d7091-fba8-4d8e-b487-20c208adc0d0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.297207] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1049.297207] env[62600]: value = "task-1223002" [ 1049.297207] env[62600]: _type = "Task" [ 1049.297207] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.304887] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223002, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.731745] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.720s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.756312] env[62600]: INFO nova.scheduler.client.report [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Deleted allocations for instance b967fb11-e70e-4e17-b769-38da581bd83b [ 1049.773100] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.773469] env[62600]: DEBUG nova.compute.manager [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Instance network_info: |[{"id": "5796bb3c-0926-47b4-bcdb-c54c18f83200", "address": "fa:16:3e:94:99:6b", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5796bb3c-09", "ovs_interfaceid": "5796bb3c-0926-47b4-bcdb-c54c18f83200", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1049.773901] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:99:6b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '184687d6-125a-4b58-bb5b-fdb404088eda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5796bb3c-0926-47b4-bcdb-c54c18f83200', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1049.781358] env[62600]: DEBUG oslo.service.loopingcall [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1049.781625] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1049.781799] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9903ee7e-e853-4163-ace6-0c2e1542ebd6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.808023] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223002, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058576} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.808719] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1049.808893] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1049.808893] env[62600]: value = "task-1223003" [ 1049.808893] env[62600]: _type = "Task" [ 1049.808893] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.809603] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a3358fe-7989-43c5-b3ec-6cb361c23084 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.837254] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 692fa608-39f4-4d7d-9f30-b6559878b970/692fa608-39f4-4d7d-9f30-b6559878b970.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1049.841999] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57dac673-e05d-4f17-bf98-8fe2a3791a25 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.857854] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1223003, 'name': CreateVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.864298] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1049.864298] env[62600]: value = "task-1223004" [ 1049.864298] env[62600]: _type = "Task" [ 1049.864298] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.875762] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223004, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.896436] env[62600]: DEBUG nova.compute.manager [req-dac2743d-0bdb-46b5-9434-46308182e269 req-aa662f56-81f2-4ca6-8909-8542381adcfe service nova] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Received event network-vif-deleted-acb612e6-f57e-4eae-8764-c9525e4c6baa {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1049.896625] env[62600]: INFO nova.compute.manager [req-dac2743d-0bdb-46b5-9434-46308182e269 req-aa662f56-81f2-4ca6-8909-8542381adcfe service nova] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Neutron deleted interface acb612e6-f57e-4eae-8764-c9525e4c6baa; detaching it from the instance and deleting it from the info cache [ 1049.896805] env[62600]: DEBUG nova.network.neutron [req-dac2743d-0bdb-46b5-9434-46308182e269 req-aa662f56-81f2-4ca6-8909-8542381adcfe service nova] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.999837] env[62600]: DEBUG nova.compute.manager [req-81ed805b-7722-42bd-a810-98a0d8893116 req-1dfd4110-bc38-4c55-9d29-779d8f5495b4 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Received event network-changed-5796bb3c-0926-47b4-bcdb-c54c18f83200 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1049.999996] env[62600]: DEBUG nova.compute.manager [req-81ed805b-7722-42bd-a810-98a0d8893116 req-1dfd4110-bc38-4c55-9d29-779d8f5495b4 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Refreshing instance network info cache due to event network-changed-5796bb3c-0926-47b4-bcdb-c54c18f83200. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1050.000291] env[62600]: DEBUG oslo_concurrency.lockutils [req-81ed805b-7722-42bd-a810-98a0d8893116 req-1dfd4110-bc38-4c55-9d29-779d8f5495b4 service nova] Acquiring lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.000489] env[62600]: DEBUG oslo_concurrency.lockutils [req-81ed805b-7722-42bd-a810-98a0d8893116 req-1dfd4110-bc38-4c55-9d29-779d8f5495b4 service nova] Acquired lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.000719] env[62600]: DEBUG nova.network.neutron [req-81ed805b-7722-42bd-a810-98a0d8893116 req-1dfd4110-bc38-4c55-9d29-779d8f5495b4 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Refreshing network info cache for port 5796bb3c-0926-47b4-bcdb-c54c18f83200 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1050.266586] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3548a931-3dc1-4cfc-9f2b-5f66c09d7dfe tempest-ServerRescueNegativeTestJSON-1498448578 tempest-ServerRescueNegativeTestJSON-1498448578-project-member] Lock "b967fb11-e70e-4e17-b769-38da581bd83b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.240s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.323153] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1223003, 'name': CreateVM_Task, 'duration_secs': 0.414865} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.323153] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1050.323727] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.323900] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.324249] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1050.324506] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9ab95d8-11ba-4cf8-99b9-128f90e9e4f9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.329277] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1050.329277] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f4cca7-d364-d225-4b23-5b90ddefd295" [ 1050.329277] env[62600]: _type = "Task" [ 1050.329277] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.336892] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f4cca7-d364-d225-4b23-5b90ddefd295, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.360020] env[62600]: DEBUG nova.network.neutron [-] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.373944] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223004, 'name': ReconfigVM_Task, 'duration_secs': 0.264784} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.375409] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 692fa608-39f4-4d7d-9f30-b6559878b970/692fa608-39f4-4d7d-9f30-b6559878b970.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1050.375409] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd595764-089e-4f51-ae1b-b9434802efdf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.382807] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1050.382807] env[62600]: value = "task-1223005" [ 1050.382807] env[62600]: _type = "Task" [ 1050.382807] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.391660] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223005, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.399399] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8a7b6761-83cf-4bca-ab04-83955d571fd6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.412974] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb05f66-c61a-4841-a35a-410d11173b92 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.441072] env[62600]: DEBUG oslo_concurrency.lockutils [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "0416ef03-fd45-405f-9047-76b34a839217" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.441414] env[62600]: DEBUG oslo_concurrency.lockutils [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "0416ef03-fd45-405f-9047-76b34a839217" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.441676] env[62600]: DEBUG oslo_concurrency.lockutils [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "0416ef03-fd45-405f-9047-76b34a839217-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.441912] env[62600]: DEBUG oslo_concurrency.lockutils [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "0416ef03-fd45-405f-9047-76b34a839217-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.442147] env[62600]: DEBUG oslo_concurrency.lockutils [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "0416ef03-fd45-405f-9047-76b34a839217-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.444227] env[62600]: DEBUG nova.compute.manager [req-dac2743d-0bdb-46b5-9434-46308182e269 req-aa662f56-81f2-4ca6-8909-8542381adcfe service nova] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Detach interface failed, port_id=acb612e6-f57e-4eae-8764-c9525e4c6baa, reason: Instance 4607dacd-28ab-4173-ba86-8a78bec19dbc could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1050.445293] env[62600]: INFO nova.compute.manager [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Terminating instance [ 1050.447267] env[62600]: DEBUG nova.compute.manager [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1050.447514] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1050.448354] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b107942-f32c-4e76-affe-ad77147651fc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.454903] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1050.455187] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7b2984f5-7144-45a2-aa77-2f83bf5c45b0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.461224] env[62600]: DEBUG oslo_vmware.api [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1050.461224] env[62600]: value = "task-1223006" [ 1050.461224] env[62600]: _type = "Task" [ 1050.461224] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.469494] env[62600]: DEBUG oslo_vmware.api [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223006, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.749995] env[62600]: DEBUG nova.network.neutron [req-81ed805b-7722-42bd-a810-98a0d8893116 req-1dfd4110-bc38-4c55-9d29-779d8f5495b4 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Updated VIF entry in instance network info cache for port 5796bb3c-0926-47b4-bcdb-c54c18f83200. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1050.754918] env[62600]: DEBUG nova.network.neutron [req-81ed805b-7722-42bd-a810-98a0d8893116 req-1dfd4110-bc38-4c55-9d29-779d8f5495b4 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Updating instance_info_cache with network_info: [{"id": "5796bb3c-0926-47b4-bcdb-c54c18f83200", "address": "fa:16:3e:94:99:6b", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5796bb3c-09", "ovs_interfaceid": "5796bb3c-0926-47b4-bcdb-c54c18f83200", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.840700] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f4cca7-d364-d225-4b23-5b90ddefd295, 'name': SearchDatastore_Task, 'duration_secs': 0.009009} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.841031] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.841275] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1050.841558] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.841678] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.841885] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1050.842187] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b65cd29-b3b5-4380-89dd-5eb497565aa6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.850476] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1050.850729] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1050.851667] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75d12da0-b2bd-40bc-ba99-9a06cba113bd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.858996] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1050.858996] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521f49c3-032d-44f6-1bff-038744fb6f76" [ 1050.858996] env[62600]: _type = "Task" [ 1050.858996] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.862580] env[62600]: INFO nova.compute.manager [-] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Took 1.68 seconds to deallocate network for instance. [ 1050.867026] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521f49c3-032d-44f6-1bff-038744fb6f76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.890520] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.890777] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.897207] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223005, 'name': Rename_Task, 'duration_secs': 0.196142} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.897483] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1050.897724] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c3a081f8-892c-4eeb-a2af-60361a723c7a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.905495] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1050.905495] env[62600]: value = "task-1223007" [ 1050.905495] env[62600]: _type = "Task" [ 1050.905495] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.918047] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223007, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.970417] env[62600]: DEBUG oslo_vmware.api [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223006, 'name': PowerOffVM_Task, 'duration_secs': 0.20815} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.970690] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1050.970868] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1050.971142] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8f3bb49b-fc9f-4640-a9ee-77f322e88aef {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.030866] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1051.031121] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1051.031318] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleting the datastore file [datastore2] 0416ef03-fd45-405f-9047-76b34a839217 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1051.031633] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-919cd0c3-b442-425c-b144-6454315b0aae {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.038693] env[62600]: DEBUG oslo_vmware.api [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1051.038693] env[62600]: value = "task-1223009" [ 1051.038693] env[62600]: _type = "Task" [ 1051.038693] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.046980] env[62600]: DEBUG oslo_vmware.api [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223009, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.253227] env[62600]: DEBUG oslo_concurrency.lockutils [req-81ed805b-7722-42bd-a810-98a0d8893116 req-1dfd4110-bc38-4c55-9d29-779d8f5495b4 service nova] Releasing lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.368423] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521f49c3-032d-44f6-1bff-038744fb6f76, 'name': SearchDatastore_Task, 'duration_secs': 0.007588} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.369210] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc3b0ef3-171e-4995-949d-d8cdaf957f3e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.372173] env[62600]: DEBUG oslo_concurrency.lockutils [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.372395] env[62600]: DEBUG oslo_concurrency.lockutils [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.373576] env[62600]: DEBUG nova.objects.instance [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lazy-loading 'resources' on Instance uuid 4607dacd-28ab-4173-ba86-8a78bec19dbc {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1051.377799] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1051.377799] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52033dbd-c7a2-2d0d-c994-2c96bfb25644" [ 1051.377799] env[62600]: _type = "Task" [ 1051.377799] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.386539] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52033dbd-c7a2-2d0d-c994-2c96bfb25644, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.393349] env[62600]: DEBUG nova.compute.manager [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1051.415438] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223007, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.550783] env[62600]: DEBUG oslo_vmware.api [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223009, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176182} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.552056] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1051.552056] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1051.552056] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1051.552056] env[62600]: INFO nova.compute.manager [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1051.552262] env[62600]: DEBUG oslo.service.loopingcall [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1051.552324] env[62600]: DEBUG nova.compute.manager [-] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1051.552458] env[62600]: DEBUG nova.network.neutron [-] [instance: 0416ef03-fd45-405f-9047-76b34a839217] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1051.888200] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52033dbd-c7a2-2d0d-c994-2c96bfb25644, 'name': SearchDatastore_Task, 'duration_secs': 0.00953} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.888629] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.888964] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 1817183b-44a5-41ec-a2bf-818edee264b3/1817183b-44a5-41ec-a2bf-818edee264b3.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1051.889289] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e13a931e-2636-43ab-837e-c59c1ec13102 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.899272] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1051.899272] env[62600]: value = "task-1223010" [ 1051.899272] env[62600]: _type = "Task" [ 1051.899272] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.917998] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223010, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.922156] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223007, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.926065] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.936048] env[62600]: DEBUG nova.compute.manager [req-51c18350-5495-44f8-9229-1a881c019410 req-a4aa5e77-d881-41f3-9b3c-b83a22f310d3 service nova] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Received event network-vif-deleted-bb15b28d-be8f-4fff-b8a6-0801d3e0541a {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1051.936048] env[62600]: INFO nova.compute.manager [req-51c18350-5495-44f8-9229-1a881c019410 req-a4aa5e77-d881-41f3-9b3c-b83a22f310d3 service nova] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Neutron deleted interface bb15b28d-be8f-4fff-b8a6-0801d3e0541a; detaching it from the instance and deleting it from the info cache [ 1051.936048] env[62600]: DEBUG nova.network.neutron [req-51c18350-5495-44f8-9229-1a881c019410 req-a4aa5e77-d881-41f3-9b3c-b83a22f310d3 service nova] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.064066] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11824444-0d0a-4a82-8052-77e2bde4c499 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.073550] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05bc1121-2527-4625-93bb-29f4cc0c60c6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.109225] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ab0d50-9936-4225-a436-2e67a3cafbe6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.119918] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-437abf54-af8f-40df-9d85-1dc0f2f77f41 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.138322] env[62600]: DEBUG nova.compute.provider_tree [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1052.296107] env[62600]: DEBUG nova.network.neutron [-] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.415401] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223010, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.420471] env[62600]: DEBUG oslo_vmware.api [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223007, 'name': PowerOnVM_Task, 'duration_secs': 1.183901} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.420728] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1052.420943] env[62600]: INFO nova.compute.manager [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Took 8.57 seconds to spawn the instance on the hypervisor. [ 1052.421150] env[62600]: DEBUG nova.compute.manager [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1052.421930] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e3b0dfa-6b12-45e8-820b-bac8bdb10680 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.443598] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d948f852-33f7-4a44-9723-0b31c7deeb32 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.452435] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87258b17-63bb-4c0d-a8fd-e40f421238ee {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.478507] env[62600]: DEBUG nova.compute.manager [req-51c18350-5495-44f8-9229-1a881c019410 req-a4aa5e77-d881-41f3-9b3c-b83a22f310d3 service nova] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Detach interface failed, port_id=bb15b28d-be8f-4fff-b8a6-0801d3e0541a, reason: Instance 0416ef03-fd45-405f-9047-76b34a839217 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1052.641843] env[62600]: DEBUG nova.scheduler.client.report [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1052.799235] env[62600]: INFO nova.compute.manager [-] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Took 1.25 seconds to deallocate network for instance. [ 1052.915488] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223010, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526166} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.915838] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] 1817183b-44a5-41ec-a2bf-818edee264b3/1817183b-44a5-41ec-a2bf-818edee264b3.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1052.915978] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1052.916263] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ed3bc22e-81f3-4c9d-bf9e-95e85afbd8a5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.923493] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1052.923493] env[62600]: value = "task-1223011" [ 1052.923493] env[62600]: _type = "Task" [ 1052.923493] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.931500] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223011, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.938692] env[62600]: INFO nova.compute.manager [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Took 15.35 seconds to build instance. [ 1053.151026] env[62600]: DEBUG oslo_concurrency.lockutils [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.776s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.151816] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.226s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.153624] env[62600]: INFO nova.compute.claims [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1053.173293] env[62600]: INFO nova.scheduler.client.report [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleted allocations for instance 4607dacd-28ab-4173-ba86-8a78bec19dbc [ 1053.306013] env[62600]: DEBUG oslo_concurrency.lockutils [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.434686] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223011, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.32262} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.434930] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1053.435852] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-528a358c-894d-4052-92cb-4e065e34207c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.442604] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0bc16d06-d6ee-46af-a6e6-542757f4113f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "692fa608-39f4-4d7d-9f30-b6559878b970" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.864s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.471467] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 1817183b-44a5-41ec-a2bf-818edee264b3/1817183b-44a5-41ec-a2bf-818edee264b3.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1053.471769] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-203d4828-07d2-4440-894a-4b151658c32e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.492102] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1053.492102] env[62600]: value = "task-1223012" [ 1053.492102] env[62600]: _type = "Task" [ 1053.492102] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.501201] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223012, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.682444] env[62600]: DEBUG oslo_concurrency.lockutils [None req-32cd17f2-f9bb-433e-845d-cf81cdb4394f tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "4607dacd-28ab-4173-ba86-8a78bec19dbc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.634s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.687185] env[62600]: DEBUG nova.compute.manager [req-efcb339c-417d-477e-af7e-11c4f9320a94 req-8a5fa605-f68b-44d1-906a-ec8d0a550c20 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Received event network-changed-280baabd-3083-4f33-b294-1f988193665b {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1053.687717] env[62600]: DEBUG nova.compute.manager [req-efcb339c-417d-477e-af7e-11c4f9320a94 req-8a5fa605-f68b-44d1-906a-ec8d0a550c20 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Refreshing instance network info cache due to event network-changed-280baabd-3083-4f33-b294-1f988193665b. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1053.687898] env[62600]: DEBUG oslo_concurrency.lockutils [req-efcb339c-417d-477e-af7e-11c4f9320a94 req-8a5fa605-f68b-44d1-906a-ec8d0a550c20 service nova] Acquiring lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.688316] env[62600]: DEBUG oslo_concurrency.lockutils [req-efcb339c-417d-477e-af7e-11c4f9320a94 req-8a5fa605-f68b-44d1-906a-ec8d0a550c20 service nova] Acquired lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.692974] env[62600]: DEBUG nova.network.neutron [req-efcb339c-417d-477e-af7e-11c4f9320a94 req-8a5fa605-f68b-44d1-906a-ec8d0a550c20 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Refreshing network info cache for port 280baabd-3083-4f33-b294-1f988193665b {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1054.001549] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223012, 'name': ReconfigVM_Task, 'duration_secs': 0.292916} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.001873] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 1817183b-44a5-41ec-a2bf-818edee264b3/1817183b-44a5-41ec-a2bf-818edee264b3.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1054.002449] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-80591ec7-3950-4217-a325-14d1ac1d3a43 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.008314] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1054.008314] env[62600]: value = "task-1223013" [ 1054.008314] env[62600]: _type = "Task" [ 1054.008314] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.016010] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223013, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.275952] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65eb4d25-1db8-4e24-8770-297f04e50353 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.285770] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe60017-3414-4be8-8142-9d3e0dd6b291 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.316749] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3494a2fc-fa9a-430a-aa0e-85e6ca4230f9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.324060] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64569758-b5c3-4e2d-b3df-34857e4033c1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.337299] env[62600]: DEBUG nova.compute.provider_tree [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1054.518499] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223013, 'name': Rename_Task, 'duration_secs': 0.163245} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.521710] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1054.522139] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3c5f8ada-5ab7-4720-a520-b81e5646bac2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.528752] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1054.528752] env[62600]: value = "task-1223014" [ 1054.528752] env[62600]: _type = "Task" [ 1054.528752] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.536155] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223014, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.586930] env[62600]: DEBUG nova.network.neutron [req-efcb339c-417d-477e-af7e-11c4f9320a94 req-8a5fa605-f68b-44d1-906a-ec8d0a550c20 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Updated VIF entry in instance network info cache for port 280baabd-3083-4f33-b294-1f988193665b. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1054.587407] env[62600]: DEBUG nova.network.neutron [req-efcb339c-417d-477e-af7e-11c4f9320a94 req-8a5fa605-f68b-44d1-906a-ec8d0a550c20 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Updating instance_info_cache with network_info: [{"id": "280baabd-3083-4f33-b294-1f988193665b", "address": "fa:16:3e:72:db:a2", "network": {"id": "dc21424f-b891-428d-a74e-76ab78d757a8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1428037886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "952323d62f034126a9b46750d03fe6ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap280baabd-30", "ovs_interfaceid": "280baabd-3083-4f33-b294-1f988193665b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.845878] env[62600]: DEBUG nova.scheduler.client.report [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1055.039021] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223014, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.090964] env[62600]: DEBUG oslo_concurrency.lockutils [req-efcb339c-417d-477e-af7e-11c4f9320a94 req-8a5fa605-f68b-44d1-906a-ec8d0a550c20 service nova] Releasing lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.311547] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "1d6bc00e-b18b-4333-b500-bb9d9d00d4e8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.311790] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "1d6bc00e-b18b-4333-b500-bb9d9d00d4e8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.352402] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.200s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.352920] env[62600]: DEBUG nova.compute.manager [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1055.355421] env[62600]: DEBUG oslo_concurrency.lockutils [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.050s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.355630] env[62600]: DEBUG oslo_concurrency.lockutils [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.374687] env[62600]: INFO nova.scheduler.client.report [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleted allocations for instance 0416ef03-fd45-405f-9047-76b34a839217 [ 1055.539499] env[62600]: DEBUG oslo_vmware.api [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223014, 'name': PowerOnVM_Task, 'duration_secs': 0.573506} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.539608] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1055.539749] env[62600]: INFO nova.compute.manager [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Took 7.75 seconds to spawn the instance on the hypervisor. [ 1055.539935] env[62600]: DEBUG nova.compute.manager [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1055.540695] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ab5aa91-7b0a-4874-b22c-89080a4b5efb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.814130] env[62600]: DEBUG nova.compute.manager [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1055.860312] env[62600]: DEBUG nova.compute.utils [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1055.862304] env[62600]: DEBUG nova.compute.manager [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1055.862482] env[62600]: DEBUG nova.network.neutron [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1055.882466] env[62600]: DEBUG oslo_concurrency.lockutils [None req-77ded4f3-8502-44ff-88c8-02e85fa3857b tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "0416ef03-fd45-405f-9047-76b34a839217" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.441s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.923654] env[62600]: DEBUG nova.policy [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bd0d91fa2ac74c38b9e0ebf39c3baa7a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73e664d178f7484a9f4741b4d9450e68', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 1056.057461] env[62600]: INFO nova.compute.manager [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Took 14.42 seconds to build instance. [ 1056.179033] env[62600]: DEBUG nova.network.neutron [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Successfully created port: 5e497e40-6c65-4c5b-aa7c-c941f82486b8 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1056.336857] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.337147] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.338831] env[62600]: INFO nova.compute.claims [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1056.362673] env[62600]: DEBUG nova.compute.manager [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1056.559388] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3b2b3897-566d-436d-b045-1cc75404685d tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "1817183b-44a5-41ec-a2bf-818edee264b3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.932s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.374310] env[62600]: DEBUG nova.compute.manager [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1057.401121] env[62600]: DEBUG nova.virt.hardware [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1057.401609] env[62600]: DEBUG nova.virt.hardware [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1057.401609] env[62600]: DEBUG nova.virt.hardware [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1057.401821] env[62600]: DEBUG nova.virt.hardware [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1057.402123] env[62600]: DEBUG nova.virt.hardware [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1057.402123] env[62600]: DEBUG nova.virt.hardware [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1057.402357] env[62600]: DEBUG nova.virt.hardware [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1057.402544] env[62600]: DEBUG nova.virt.hardware [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1057.402729] env[62600]: DEBUG nova.virt.hardware [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1057.402900] env[62600]: DEBUG nova.virt.hardware [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1057.403096] env[62600]: DEBUG nova.virt.hardware [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1057.404306] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c4250b-581e-435b-9033-31899ffa5200 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.413537] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e16dd3-e72a-4cf0-84d9-fa0a34e74d88 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.473420] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c5b158-2398-4750-bd27-522751141ff9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.480843] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26fd9fc7-c2b2-4285-9bfb-2a493a250f27 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.512645] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31118728-c508-43e4-b834-482f9f9b5e5e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.517662] env[62600]: DEBUG nova.compute.manager [req-277daf03-a8e8-4493-be10-db57b9a0aa3d req-3780e628-3b5f-4f25-a109-fdb6d2f691e1 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Received event network-changed-b8e9cd2e-1990-4997-b1fa-cb4411d39fde {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1057.517662] env[62600]: DEBUG nova.compute.manager [req-277daf03-a8e8-4493-be10-db57b9a0aa3d req-3780e628-3b5f-4f25-a109-fdb6d2f691e1 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Refreshing instance network info cache due to event network-changed-b8e9cd2e-1990-4997-b1fa-cb4411d39fde. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1057.517910] env[62600]: DEBUG oslo_concurrency.lockutils [req-277daf03-a8e8-4493-be10-db57b9a0aa3d req-3780e628-3b5f-4f25-a109-fdb6d2f691e1 service nova] Acquiring lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.517910] env[62600]: DEBUG oslo_concurrency.lockutils [req-277daf03-a8e8-4493-be10-db57b9a0aa3d req-3780e628-3b5f-4f25-a109-fdb6d2f691e1 service nova] Acquired lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.518018] env[62600]: DEBUG nova.network.neutron [req-277daf03-a8e8-4493-be10-db57b9a0aa3d req-3780e628-3b5f-4f25-a109-fdb6d2f691e1 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Refreshing network info cache for port b8e9cd2e-1990-4997-b1fa-cb4411d39fde {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1057.525148] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cae33f7e-fbce-48e4-a8c8-c8b05447f3cb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.539938] env[62600]: DEBUG nova.compute.provider_tree [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1057.698065] env[62600]: DEBUG nova.network.neutron [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Successfully updated port: 5e497e40-6c65-4c5b-aa7c-c941f82486b8 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1058.042710] env[62600]: DEBUG nova.scheduler.client.report [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1058.200394] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "refresh_cache-9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.200550] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "refresh_cache-9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.200696] env[62600]: DEBUG nova.network.neutron [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1058.379211] env[62600]: DEBUG nova.network.neutron [req-277daf03-a8e8-4493-be10-db57b9a0aa3d req-3780e628-3b5f-4f25-a109-fdb6d2f691e1 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Updated VIF entry in instance network info cache for port b8e9cd2e-1990-4997-b1fa-cb4411d39fde. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1058.379599] env[62600]: DEBUG nova.network.neutron [req-277daf03-a8e8-4493-be10-db57b9a0aa3d req-3780e628-3b5f-4f25-a109-fdb6d2f691e1 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Updating instance_info_cache with network_info: [{"id": "b8e9cd2e-1990-4997-b1fa-cb4411d39fde", "address": "fa:16:3e:fc:c1:ac", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8e9cd2e-19", "ovs_interfaceid": "b8e9cd2e-1990-4997-b1fa-cb4411d39fde", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.547620] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.210s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.548174] env[62600]: DEBUG nova.compute.manager [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1058.782436] env[62600]: DEBUG nova.network.neutron [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1058.882150] env[62600]: DEBUG oslo_concurrency.lockutils [req-277daf03-a8e8-4493-be10-db57b9a0aa3d req-3780e628-3b5f-4f25-a109-fdb6d2f691e1 service nova] Releasing lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.990531] env[62600]: DEBUG nova.network.neutron [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Updating instance_info_cache with network_info: [{"id": "5e497e40-6c65-4c5b-aa7c-c941f82486b8", "address": "fa:16:3e:cb:ec:95", "network": {"id": "95818650-8227-4d11-9d14-007f06c07790", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1738152115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73e664d178f7484a9f4741b4d9450e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e497e40-6c", "ovs_interfaceid": "5e497e40-6c65-4c5b-aa7c-c941f82486b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.055357] env[62600]: DEBUG nova.compute.utils [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1059.055357] env[62600]: DEBUG nova.compute.manager [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1059.055357] env[62600]: DEBUG nova.network.neutron [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1059.151129] env[62600]: DEBUG nova.policy [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b3e59998217a4b18a6f3f01142a5e440', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f91091f83ee4a2091507ca994e3d52f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 1059.380187] env[62600]: DEBUG nova.network.neutron [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Successfully created port: c39361b1-8fba-41ad-a28f-a72eeafde1e2 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1059.494228] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "refresh_cache-9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.494530] env[62600]: DEBUG nova.compute.manager [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Instance network_info: |[{"id": "5e497e40-6c65-4c5b-aa7c-c941f82486b8", "address": "fa:16:3e:cb:ec:95", "network": {"id": "95818650-8227-4d11-9d14-007f06c07790", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1738152115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73e664d178f7484a9f4741b4d9450e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e497e40-6c", "ovs_interfaceid": "5e497e40-6c65-4c5b-aa7c-c941f82486b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1059.494989] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cb:ec:95', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43798f54-0c07-4417-a23f-58bb6b7e204b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5e497e40-6c65-4c5b-aa7c-c941f82486b8', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1059.502388] env[62600]: DEBUG oslo.service.loopingcall [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1059.502828] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1059.503716] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-02245682-e135-4951-bcc7-05fdaff605dc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.523746] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1059.523746] env[62600]: value = "task-1223015" [ 1059.523746] env[62600]: _type = "Task" [ 1059.523746] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.533564] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1223015, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.560102] env[62600]: DEBUG nova.compute.manager [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Received event network-changed-5796bb3c-0926-47b4-bcdb-c54c18f83200 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1059.560102] env[62600]: DEBUG nova.compute.manager [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Refreshing instance network info cache due to event network-changed-5796bb3c-0926-47b4-bcdb-c54c18f83200. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1059.560102] env[62600]: DEBUG oslo_concurrency.lockutils [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] Acquiring lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.560102] env[62600]: DEBUG oslo_concurrency.lockutils [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] Acquired lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.560102] env[62600]: DEBUG nova.network.neutron [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Refreshing network info cache for port 5796bb3c-0926-47b4-bcdb-c54c18f83200 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1059.560102] env[62600]: DEBUG nova.compute.manager [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1060.034496] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1223015, 'name': CreateVM_Task, 'duration_secs': 0.346269} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.034669] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1060.035493] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.035663] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.035978] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1060.036265] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61410ae6-cb05-446c-bd38-de72f8730bd5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.040785] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1060.040785] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526dee95-e9de-4caf-c412-4f6f1456b35c" [ 1060.040785] env[62600]: _type = "Task" [ 1060.040785] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.048154] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526dee95-e9de-4caf-c412-4f6f1456b35c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.252952] env[62600]: DEBUG nova.network.neutron [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Updated VIF entry in instance network info cache for port 5796bb3c-0926-47b4-bcdb-c54c18f83200. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1060.253409] env[62600]: DEBUG nova.network.neutron [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Updating instance_info_cache with network_info: [{"id": "5796bb3c-0926-47b4-bcdb-c54c18f83200", "address": "fa:16:3e:94:99:6b", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5796bb3c-09", "ovs_interfaceid": "5796bb3c-0926-47b4-bcdb-c54c18f83200", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.551379] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526dee95-e9de-4caf-c412-4f6f1456b35c, 'name': SearchDatastore_Task, 'duration_secs': 0.009022} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.551713] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.551947] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1060.552271] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.552467] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.552696] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1060.553000] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8a55746f-747e-4dce-a56a-dea7fe5994a8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.560966] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1060.561155] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1060.561828] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5da5dca4-6aa5-4828-ae47-6dae8d9ed20f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.566682] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1060.566682] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c9bc31-a53a-fbd3-8d33-5c1bbf5042ba" [ 1060.566682] env[62600]: _type = "Task" [ 1060.566682] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.570812] env[62600]: DEBUG nova.compute.manager [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1060.577703] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c9bc31-a53a-fbd3-8d33-5c1bbf5042ba, 'name': SearchDatastore_Task, 'duration_secs': 0.007652} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.578435] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2c5baaa-7bf1-4625-9b8a-e6d3df7d90af {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.583031] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1060.583031] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5237215e-2b1e-c0ee-496f-41bce5c0ceba" [ 1060.583031] env[62600]: _type = "Task" [ 1060.583031] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.591053] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5237215e-2b1e-c0ee-496f-41bce5c0ceba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.598231] env[62600]: DEBUG nova.virt.hardware [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1060.598471] env[62600]: DEBUG nova.virt.hardware [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1060.598636] env[62600]: DEBUG nova.virt.hardware [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1060.598825] env[62600]: DEBUG nova.virt.hardware [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1060.598981] env[62600]: DEBUG nova.virt.hardware [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1060.599155] env[62600]: DEBUG nova.virt.hardware [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1060.599379] env[62600]: DEBUG nova.virt.hardware [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1060.599544] env[62600]: DEBUG nova.virt.hardware [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1060.599719] env[62600]: DEBUG nova.virt.hardware [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1060.599890] env[62600]: DEBUG nova.virt.hardware [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1060.600085] env[62600]: DEBUG nova.virt.hardware [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1060.600818] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d88ba6e6-37f3-489a-92e3-584cd173c3fc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.607793] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8905a8a2-bc98-4927-98e2-1f4313c58e9f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.756301] env[62600]: DEBUG oslo_concurrency.lockutils [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] Releasing lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.756599] env[62600]: DEBUG nova.compute.manager [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Received event network-vif-plugged-5e497e40-6c65-4c5b-aa7c-c941f82486b8 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1060.756809] env[62600]: DEBUG oslo_concurrency.lockutils [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] Acquiring lock "9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.757041] env[62600]: DEBUG oslo_concurrency.lockutils [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] Lock "9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.757218] env[62600]: DEBUG oslo_concurrency.lockutils [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] Lock "9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.757395] env[62600]: DEBUG nova.compute.manager [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] No waiting events found dispatching network-vif-plugged-5e497e40-6c65-4c5b-aa7c-c941f82486b8 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1060.757589] env[62600]: WARNING nova.compute.manager [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Received unexpected event network-vif-plugged-5e497e40-6c65-4c5b-aa7c-c941f82486b8 for instance with vm_state building and task_state spawning. [ 1060.757771] env[62600]: DEBUG nova.compute.manager [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Received event network-changed-5e497e40-6c65-4c5b-aa7c-c941f82486b8 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1060.757977] env[62600]: DEBUG nova.compute.manager [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Refreshing instance network info cache due to event network-changed-5e497e40-6c65-4c5b-aa7c-c941f82486b8. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1060.758141] env[62600]: DEBUG oslo_concurrency.lockutils [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] Acquiring lock "refresh_cache-9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.758287] env[62600]: DEBUG oslo_concurrency.lockutils [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] Acquired lock "refresh_cache-9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.758451] env[62600]: DEBUG nova.network.neutron [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Refreshing network info cache for port 5e497e40-6c65-4c5b-aa7c-c941f82486b8 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1060.840660] env[62600]: DEBUG nova.network.neutron [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Successfully updated port: c39361b1-8fba-41ad-a28f-a72eeafde1e2 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1061.096329] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5237215e-2b1e-c0ee-496f-41bce5c0ceba, 'name': SearchDatastore_Task, 'duration_secs': 0.007729} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.096329] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.096501] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7/9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1061.096796] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d216ed3b-1653-4c40-993d-c17d0434677d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.105845] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1061.105845] env[62600]: value = "task-1223016" [ 1061.105845] env[62600]: _type = "Task" [ 1061.105845] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.114276] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223016, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.343594] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "refresh_cache-1d6bc00e-b18b-4333-b500-bb9d9d00d4e8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.343720] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "refresh_cache-1d6bc00e-b18b-4333-b500-bb9d9d00d4e8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.343873] env[62600]: DEBUG nova.network.neutron [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1061.563743] env[62600]: DEBUG nova.network.neutron [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Updated VIF entry in instance network info cache for port 5e497e40-6c65-4c5b-aa7c-c941f82486b8. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1061.564153] env[62600]: DEBUG nova.network.neutron [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Updating instance_info_cache with network_info: [{"id": "5e497e40-6c65-4c5b-aa7c-c941f82486b8", "address": "fa:16:3e:cb:ec:95", "network": {"id": "95818650-8227-4d11-9d14-007f06c07790", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1738152115-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73e664d178f7484a9f4741b4d9450e68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e497e40-6c", "ovs_interfaceid": "5e497e40-6c65-4c5b-aa7c-c941f82486b8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.580811] env[62600]: DEBUG nova.compute.manager [req-9ba02fa5-1a50-43de-a9f1-5caa1c542bab req-05132a40-4419-44e4-9614-2774047046a7 service nova] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Received event network-vif-plugged-c39361b1-8fba-41ad-a28f-a72eeafde1e2 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1061.581062] env[62600]: DEBUG oslo_concurrency.lockutils [req-9ba02fa5-1a50-43de-a9f1-5caa1c542bab req-05132a40-4419-44e4-9614-2774047046a7 service nova] Acquiring lock "1d6bc00e-b18b-4333-b500-bb9d9d00d4e8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.581292] env[62600]: DEBUG oslo_concurrency.lockutils [req-9ba02fa5-1a50-43de-a9f1-5caa1c542bab req-05132a40-4419-44e4-9614-2774047046a7 service nova] Lock "1d6bc00e-b18b-4333-b500-bb9d9d00d4e8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.581452] env[62600]: DEBUG oslo_concurrency.lockutils [req-9ba02fa5-1a50-43de-a9f1-5caa1c542bab req-05132a40-4419-44e4-9614-2774047046a7 service nova] Lock "1d6bc00e-b18b-4333-b500-bb9d9d00d4e8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.581667] env[62600]: DEBUG nova.compute.manager [req-9ba02fa5-1a50-43de-a9f1-5caa1c542bab req-05132a40-4419-44e4-9614-2774047046a7 service nova] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] No waiting events found dispatching network-vif-plugged-c39361b1-8fba-41ad-a28f-a72eeafde1e2 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1061.581791] env[62600]: WARNING nova.compute.manager [req-9ba02fa5-1a50-43de-a9f1-5caa1c542bab req-05132a40-4419-44e4-9614-2774047046a7 service nova] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Received unexpected event network-vif-plugged-c39361b1-8fba-41ad-a28f-a72eeafde1e2 for instance with vm_state building and task_state spawning. [ 1061.581955] env[62600]: DEBUG nova.compute.manager [req-9ba02fa5-1a50-43de-a9f1-5caa1c542bab req-05132a40-4419-44e4-9614-2774047046a7 service nova] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Received event network-changed-c39361b1-8fba-41ad-a28f-a72eeafde1e2 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1061.582132] env[62600]: DEBUG nova.compute.manager [req-9ba02fa5-1a50-43de-a9f1-5caa1c542bab req-05132a40-4419-44e4-9614-2774047046a7 service nova] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Refreshing instance network info cache due to event network-changed-c39361b1-8fba-41ad-a28f-a72eeafde1e2. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1061.582302] env[62600]: DEBUG oslo_concurrency.lockutils [req-9ba02fa5-1a50-43de-a9f1-5caa1c542bab req-05132a40-4419-44e4-9614-2774047046a7 service nova] Acquiring lock "refresh_cache-1d6bc00e-b18b-4333-b500-bb9d9d00d4e8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.616591] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223016, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457177} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.616796] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7/9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1061.617014] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1061.617294] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2397c724-8a0e-4323-ba1a-a0bc8cc1ffc2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.623253] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1061.623253] env[62600]: value = "task-1223017" [ 1061.623253] env[62600]: _type = "Task" [ 1061.623253] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.630106] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223017, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.873982] env[62600]: DEBUG nova.network.neutron [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1061.992480] env[62600]: DEBUG nova.network.neutron [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Updating instance_info_cache with network_info: [{"id": "c39361b1-8fba-41ad-a28f-a72eeafde1e2", "address": "fa:16:3e:37:77:2c", "network": {"id": "1d2052cf-5be1-48e7-887a-18f367b7823f", "bridge": "br-int", "label": "tempest-ServersTestJSON-106099679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f91091f83ee4a2091507ca994e3d52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc39361b1-8f", "ovs_interfaceid": "c39361b1-8fba-41ad-a28f-a72eeafde1e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.066588] env[62600]: DEBUG oslo_concurrency.lockutils [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] Releasing lock "refresh_cache-9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.066881] env[62600]: DEBUG nova.compute.manager [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Received event network-changed-5796bb3c-0926-47b4-bcdb-c54c18f83200 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1062.067071] env[62600]: DEBUG nova.compute.manager [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Refreshing instance network info cache due to event network-changed-5796bb3c-0926-47b4-bcdb-c54c18f83200. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1062.067292] env[62600]: DEBUG oslo_concurrency.lockutils [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] Acquiring lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.067462] env[62600]: DEBUG oslo_concurrency.lockutils [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] Acquired lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.067643] env[62600]: DEBUG nova.network.neutron [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Refreshing network info cache for port 5796bb3c-0926-47b4-bcdb-c54c18f83200 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1062.132827] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223017, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059162} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.133128] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1062.133896] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49bd12ec-fb51-45e9-8656-de5bf845bc11 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.155097] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7/9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1062.155332] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3fac6fea-9894-49b8-9083-baf90321d01e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.173456] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1062.173456] env[62600]: value = "task-1223018" [ 1062.173456] env[62600]: _type = "Task" [ 1062.173456] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.180632] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223018, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.494776] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "refresh_cache-1d6bc00e-b18b-4333-b500-bb9d9d00d4e8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.495300] env[62600]: DEBUG nova.compute.manager [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Instance network_info: |[{"id": "c39361b1-8fba-41ad-a28f-a72eeafde1e2", "address": "fa:16:3e:37:77:2c", "network": {"id": "1d2052cf-5be1-48e7-887a-18f367b7823f", "bridge": "br-int", "label": "tempest-ServersTestJSON-106099679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f91091f83ee4a2091507ca994e3d52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc39361b1-8f", "ovs_interfaceid": "c39361b1-8fba-41ad-a28f-a72eeafde1e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1062.495695] env[62600]: DEBUG oslo_concurrency.lockutils [req-9ba02fa5-1a50-43de-a9f1-5caa1c542bab req-05132a40-4419-44e4-9614-2774047046a7 service nova] Acquired lock "refresh_cache-1d6bc00e-b18b-4333-b500-bb9d9d00d4e8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.495920] env[62600]: DEBUG nova.network.neutron [req-9ba02fa5-1a50-43de-a9f1-5caa1c542bab req-05132a40-4419-44e4-9614-2774047046a7 service nova] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Refreshing network info cache for port c39361b1-8fba-41ad-a28f-a72eeafde1e2 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1062.497262] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:37:77:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '028bae2d-fe6c-4207-b4a3-3fab45fbf1d6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c39361b1-8fba-41ad-a28f-a72eeafde1e2', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1062.504743] env[62600]: DEBUG oslo.service.loopingcall [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1062.505106] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1062.505334] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6c4f2169-26ba-4759-9bd2-10b59eef215e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.526652] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1062.526652] env[62600]: value = "task-1223019" [ 1062.526652] env[62600]: _type = "Task" [ 1062.526652] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.534376] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1223019, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.683385] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223018, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.759623] env[62600]: DEBUG nova.network.neutron [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Updated VIF entry in instance network info cache for port 5796bb3c-0926-47b4-bcdb-c54c18f83200. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1062.759996] env[62600]: DEBUG nova.network.neutron [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Updating instance_info_cache with network_info: [{"id": "5796bb3c-0926-47b4-bcdb-c54c18f83200", "address": "fa:16:3e:94:99:6b", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5796bb3c-09", "ovs_interfaceid": "5796bb3c-0926-47b4-bcdb-c54c18f83200", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.038985] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1223019, 'name': CreateVM_Task} progress is 25%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.183485] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223018, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.191736] env[62600]: DEBUG nova.network.neutron [req-9ba02fa5-1a50-43de-a9f1-5caa1c542bab req-05132a40-4419-44e4-9614-2774047046a7 service nova] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Updated VIF entry in instance network info cache for port c39361b1-8fba-41ad-a28f-a72eeafde1e2. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1063.192116] env[62600]: DEBUG nova.network.neutron [req-9ba02fa5-1a50-43de-a9f1-5caa1c542bab req-05132a40-4419-44e4-9614-2774047046a7 service nova] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Updating instance_info_cache with network_info: [{"id": "c39361b1-8fba-41ad-a28f-a72eeafde1e2", "address": "fa:16:3e:37:77:2c", "network": {"id": "1d2052cf-5be1-48e7-887a-18f367b7823f", "bridge": "br-int", "label": "tempest-ServersTestJSON-106099679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8f91091f83ee4a2091507ca994e3d52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc39361b1-8f", "ovs_interfaceid": "c39361b1-8fba-41ad-a28f-a72eeafde1e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.262961] env[62600]: DEBUG oslo_concurrency.lockutils [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] Releasing lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.263264] env[62600]: DEBUG nova.compute.manager [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Received event network-changed-b8e9cd2e-1990-4997-b1fa-cb4411d39fde {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1063.263445] env[62600]: DEBUG nova.compute.manager [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Refreshing instance network info cache due to event network-changed-b8e9cd2e-1990-4997-b1fa-cb4411d39fde. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1063.263664] env[62600]: DEBUG oslo_concurrency.lockutils [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] Acquiring lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.263814] env[62600]: DEBUG oslo_concurrency.lockutils [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] Acquired lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.263982] env[62600]: DEBUG nova.network.neutron [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Refreshing network info cache for port b8e9cd2e-1990-4997-b1fa-cb4411d39fde {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1063.537660] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1223019, 'name': CreateVM_Task} progress is 25%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.683527] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223018, 'name': ReconfigVM_Task, 'duration_secs': 1.280908} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.683831] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7/9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1063.684436] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4504e54d-cd7d-472a-bf5c-3c18955be979 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.690328] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1063.690328] env[62600]: value = "task-1223020" [ 1063.690328] env[62600]: _type = "Task" [ 1063.690328] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.694719] env[62600]: DEBUG oslo_concurrency.lockutils [req-9ba02fa5-1a50-43de-a9f1-5caa1c542bab req-05132a40-4419-44e4-9614-2774047046a7 service nova] Releasing lock "refresh_cache-1d6bc00e-b18b-4333-b500-bb9d9d00d4e8" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.697825] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223020, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.963617] env[62600]: DEBUG nova.network.neutron [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Updated VIF entry in instance network info cache for port b8e9cd2e-1990-4997-b1fa-cb4411d39fde. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1063.964008] env[62600]: DEBUG nova.network.neutron [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Updating instance_info_cache with network_info: [{"id": "b8e9cd2e-1990-4997-b1fa-cb4411d39fde", "address": "fa:16:3e:fc:c1:ac", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8e9cd2e-19", "ovs_interfaceid": "b8e9cd2e-1990-4997-b1fa-cb4411d39fde", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.036881] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1223019, 'name': CreateVM_Task, 'duration_secs': 1.108612} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.037125] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1064.037818] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.037991] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.038346] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1064.038601] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f796ae7-c6f4-4f56-872a-73b9bf2a1ef3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.043067] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1064.043067] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d12024-6e77-923f-05ab-8b72a7058875" [ 1064.043067] env[62600]: _type = "Task" [ 1064.043067] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.050767] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d12024-6e77-923f-05ab-8b72a7058875, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.199813] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223020, 'name': Rename_Task, 'duration_secs': 0.166939} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.200130] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1064.200390] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8676d500-5d79-4a46-abd1-cdbced6292f0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.206281] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1064.206281] env[62600]: value = "task-1223021" [ 1064.206281] env[62600]: _type = "Task" [ 1064.206281] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.213467] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223021, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.467357] env[62600]: DEBUG oslo_concurrency.lockutils [req-d122526f-3883-416c-8576-fab138412e7a req-adc338b3-afea-47f8-ab73-8ae4d4ddae09 service nova] Releasing lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.554166] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d12024-6e77-923f-05ab-8b72a7058875, 'name': SearchDatastore_Task, 'duration_secs': 0.009966} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.554513] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.554801] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1064.555103] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.555302] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.555526] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1064.555819] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f9659b8d-ccc4-4b8e-b55e-6dddc9939580 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.564035] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1064.564035] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1064.564709] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bcdae0c9-2d82-4379-86aa-8d6e3432e0dc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.569808] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1064.569808] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c4623d-c5e0-703a-c5da-b444d174c657" [ 1064.569808] env[62600]: _type = "Task" [ 1064.569808] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.579098] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c4623d-c5e0-703a-c5da-b444d174c657, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.716135] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223021, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.080132] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c4623d-c5e0-703a-c5da-b444d174c657, 'name': SearchDatastore_Task, 'duration_secs': 0.008951} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.080905] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34d242c4-ac8c-478e-8627-e93892d5ff39 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.086350] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1065.086350] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5233626b-0711-38e5-a30b-5b4513685166" [ 1065.086350] env[62600]: _type = "Task" [ 1065.086350] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.093699] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5233626b-0711-38e5-a30b-5b4513685166, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.216282] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223021, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.599700] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5233626b-0711-38e5-a30b-5b4513685166, 'name': SearchDatastore_Task, 'duration_secs': 0.008592} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.600096] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.600465] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8/1d6bc00e-b18b-4333-b500-bb9d9d00d4e8.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1065.600812] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3ce76a81-ea25-4bef-b7dd-5f3fdabed513 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.608416] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1065.608416] env[62600]: value = "task-1223022" [ 1065.608416] env[62600]: _type = "Task" [ 1065.608416] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.618185] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223022, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.716926] env[62600]: DEBUG oslo_vmware.api [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223021, 'name': PowerOnVM_Task, 'duration_secs': 1.055725} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.717236] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1065.717481] env[62600]: INFO nova.compute.manager [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Took 8.34 seconds to spawn the instance on the hypervisor. [ 1065.717682] env[62600]: DEBUG nova.compute.manager [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1065.718466] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b297d0d5-0a15-4768-9f45-52c9180e1d75 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.118927] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223022, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.437416} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.119381] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8/1d6bc00e-b18b-4333-b500-bb9d9d00d4e8.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1066.119644] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1066.119935] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c12e3336-bb85-4d59-b90f-c617d0013825 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.126342] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1066.126342] env[62600]: value = "task-1223023" [ 1066.126342] env[62600]: _type = "Task" [ 1066.126342] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.134496] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223023, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.240625] env[62600]: INFO nova.compute.manager [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Took 14.34 seconds to build instance. [ 1066.636343] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223023, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064681} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.636620] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1066.637387] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcef53b2-0d61-4a1b-84dc-142f942b13a4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.658739] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8/1d6bc00e-b18b-4333-b500-bb9d9d00d4e8.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1066.659233] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0ee04b7-a236-4e40-8767-682825b83dd5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.680087] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1066.680087] env[62600]: value = "task-1223024" [ 1066.680087] env[62600]: _type = "Task" [ 1066.680087] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.687907] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223024, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.742980] env[62600]: DEBUG oslo_concurrency.lockutils [None req-b8d4e129-d5cc-4d2b-85d3-9826c8054f27 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.852s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.111506] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.111906] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.112181] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.112400] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.112796] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.115147] env[62600]: INFO nova.compute.manager [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Terminating instance [ 1067.117023] env[62600]: DEBUG nova.compute.manager [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1067.117235] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1067.118067] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9c54bb2-20ee-4d12-b420-4b38b7b9bc16 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.125608] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1067.125846] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-49655385-f90a-424c-b70b-885eecfd3609 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.131477] env[62600]: DEBUG oslo_vmware.api [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1067.131477] env[62600]: value = "task-1223025" [ 1067.131477] env[62600]: _type = "Task" [ 1067.131477] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.139244] env[62600]: DEBUG oslo_vmware.api [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223025, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.190109] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223024, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.642376] env[62600]: DEBUG oslo_vmware.api [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223025, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.690394] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223024, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.142486] env[62600]: DEBUG oslo_vmware.api [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223025, 'name': PowerOffVM_Task, 'duration_secs': 0.687891} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.142791] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1068.142951] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1068.143249] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7a0d0c5c-2a06-4b36-8c9d-0fbaf29e6b49 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.191085] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223024, 'name': ReconfigVM_Task, 'duration_secs': 1.164229} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.191085] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8/1d6bc00e-b18b-4333-b500-bb9d9d00d4e8.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1068.191632] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-206e5091-6939-4aa7-b144-61b41b73202b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.197140] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1068.197140] env[62600]: value = "task-1223027" [ 1068.197140] env[62600]: _type = "Task" [ 1068.197140] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.206104] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223027, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.219809] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1068.220043] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1068.220238] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleting the datastore file [datastore2] 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1068.220509] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-23fb663c-8c02-4184-946b-1bdbc940fbd7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.226113] env[62600]: DEBUG oslo_vmware.api [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for the task: (returnval){ [ 1068.226113] env[62600]: value = "task-1223028" [ 1068.226113] env[62600]: _type = "Task" [ 1068.226113] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.235963] env[62600]: DEBUG oslo_vmware.api [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223028, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.707943] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223027, 'name': Rename_Task, 'duration_secs': 0.163206} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.708314] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1068.708607] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c0995931-89b1-44ab-ba79-12e0d49bbe68 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.714932] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1068.714932] env[62600]: value = "task-1223029" [ 1068.714932] env[62600]: _type = "Task" [ 1068.714932] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.721886] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223029, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.733876] env[62600]: DEBUG oslo_vmware.api [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Task: {'id': task-1223028, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.124861} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.734124] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1068.734331] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1068.734566] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1068.734760] env[62600]: INFO nova.compute.manager [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Took 1.62 seconds to destroy the instance on the hypervisor. [ 1068.735054] env[62600]: DEBUG oslo.service.loopingcall [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1068.735272] env[62600]: DEBUG nova.compute.manager [-] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1068.735369] env[62600]: DEBUG nova.network.neutron [-] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1068.956339] env[62600]: DEBUG nova.compute.manager [req-84331240-7ca4-4d64-a92d-f320de0cd139 req-554c4f28-b6fc-41da-8fb8-b17387f1ffbf service nova] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Received event network-vif-deleted-5e497e40-6c65-4c5b-aa7c-c941f82486b8 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1068.956591] env[62600]: INFO nova.compute.manager [req-84331240-7ca4-4d64-a92d-f320de0cd139 req-554c4f28-b6fc-41da-8fb8-b17387f1ffbf service nova] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Neutron deleted interface 5e497e40-6c65-4c5b-aa7c-c941f82486b8; detaching it from the instance and deleting it from the info cache [ 1068.956737] env[62600]: DEBUG nova.network.neutron [req-84331240-7ca4-4d64-a92d-f320de0cd139 req-554c4f28-b6fc-41da-8fb8-b17387f1ffbf service nova] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.225064] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223029, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.436312] env[62600]: DEBUG nova.network.neutron [-] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.459855] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eabbb620-67fb-4cbb-a3a6-cc36bdbb718c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.469174] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cf16ad1-4d4d-4a6d-b17b-334f5b435757 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.495703] env[62600]: DEBUG nova.compute.manager [req-84331240-7ca4-4d64-a92d-f320de0cd139 req-554c4f28-b6fc-41da-8fb8-b17387f1ffbf service nova] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Detach interface failed, port_id=5e497e40-6c65-4c5b-aa7c-c941f82486b8, reason: Instance 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1069.724693] env[62600]: DEBUG oslo_vmware.api [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223029, 'name': PowerOnVM_Task, 'duration_secs': 0.543733} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.725026] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1069.725263] env[62600]: INFO nova.compute.manager [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Took 9.15 seconds to spawn the instance on the hypervisor. [ 1069.725465] env[62600]: DEBUG nova.compute.manager [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1069.726230] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca3e19e0-fc0b-453d-91a3-da2612dec0f7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.940060] env[62600]: INFO nova.compute.manager [-] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Took 1.20 seconds to deallocate network for instance. [ 1070.246633] env[62600]: INFO nova.compute.manager [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Took 13.93 seconds to build instance. [ 1070.337698] env[62600]: DEBUG oslo_concurrency.lockutils [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "interface-eb3c5624-0599-4e81-b056-bedf34329646-1e67e320-f1e0-40f7-b0df-9cac23dbf102" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.337969] env[62600]: DEBUG oslo_concurrency.lockutils [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "interface-eb3c5624-0599-4e81-b056-bedf34329646-1e67e320-f1e0-40f7-b0df-9cac23dbf102" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.338487] env[62600]: DEBUG nova.objects.instance [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lazy-loading 'flavor' on Instance uuid eb3c5624-0599-4e81-b056-bedf34329646 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.445455] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.445743] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.445984] env[62600]: DEBUG nova.objects.instance [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lazy-loading 'resources' on Instance uuid 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.748366] env[62600]: DEBUG oslo_concurrency.lockutils [None req-cff8e059-a6f3-419e-ab9b-c055bf81e4a0 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "1d6bc00e-b18b-4333-b500-bb9d9d00d4e8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.436s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.928143] env[62600]: DEBUG nova.objects.instance [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lazy-loading 'pci_requests' on Instance uuid eb3c5624-0599-4e81-b056-bedf34329646 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.043984] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3d0405e-4ce9-4a4a-8626-a6a00b8fc33c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.051646] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132804e7-dbdc-4bb9-8029-6b8dcd1b84a1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.081386] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a230cf17-9b02-4519-a78e-3e51aa0fd8c7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.087850] env[62600]: DEBUG oslo_concurrency.lockutils [None req-de24eb6d-3a35-4425-9b50-aeaa6096b019 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "1d6bc00e-b18b-4333-b500-bb9d9d00d4e8" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.088279] env[62600]: DEBUG oslo_concurrency.lockutils [None req-de24eb6d-3a35-4425-9b50-aeaa6096b019 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "1d6bc00e-b18b-4333-b500-bb9d9d00d4e8" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.088357] env[62600]: DEBUG nova.compute.manager [None req-de24eb6d-3a35-4425-9b50-aeaa6096b019 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1071.089067] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03674733-3d79-4c13-953e-0269beac673e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.092576] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-806ce13e-58cf-421e-befe-1c7ba3edb3f5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.099446] env[62600]: DEBUG nova.compute.manager [None req-de24eb6d-3a35-4425-9b50-aeaa6096b019 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62600) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1071.099982] env[62600]: DEBUG nova.objects.instance [None req-de24eb6d-3a35-4425-9b50-aeaa6096b019 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lazy-loading 'flavor' on Instance uuid 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.109140] env[62600]: DEBUG nova.compute.provider_tree [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.431367] env[62600]: DEBUG nova.objects.base [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62600) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1071.431710] env[62600]: DEBUG nova.network.neutron [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1071.494039] env[62600]: DEBUG nova.policy [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fbd1b2fb34d841359ada8fc44bec2986', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b028450f2da445fb83e37adfc86bba68', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 1071.613866] env[62600]: DEBUG nova.scheduler.client.report [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1071.617138] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-de24eb6d-3a35-4425-9b50-aeaa6096b019 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1071.617389] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7edb40e3-133f-4d86-af6a-bf4fffcb6612 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.624376] env[62600]: DEBUG oslo_vmware.api [None req-de24eb6d-3a35-4425-9b50-aeaa6096b019 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1071.624376] env[62600]: value = "task-1223030" [ 1071.624376] env[62600]: _type = "Task" [ 1071.624376] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.633199] env[62600]: DEBUG oslo_vmware.api [None req-de24eb6d-3a35-4425-9b50-aeaa6096b019 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223030, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.120662] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.675s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.134173] env[62600]: DEBUG oslo_vmware.api [None req-de24eb6d-3a35-4425-9b50-aeaa6096b019 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223030, 'name': PowerOffVM_Task, 'duration_secs': 0.186266} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.134405] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-de24eb6d-3a35-4425-9b50-aeaa6096b019 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1072.134688] env[62600]: DEBUG nova.compute.manager [None req-de24eb6d-3a35-4425-9b50-aeaa6096b019 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1072.135401] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e8d866-1de0-438f-b308-5c568367c9dd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.143242] env[62600]: INFO nova.scheduler.client.report [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Deleted allocations for instance 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7 [ 1072.648018] env[62600]: DEBUG oslo_concurrency.lockutils [None req-de24eb6d-3a35-4425-9b50-aeaa6096b019 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "1d6bc00e-b18b-4333-b500-bb9d9d00d4e8" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.560s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.651048] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f3c9b8f2-5d13-434a-a0d4-2ac98dfc8d94 tempest-ServerDiskConfigTestJSON-2050657305 tempest-ServerDiskConfigTestJSON-2050657305-project-member] Lock "9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.539s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.963009] env[62600]: DEBUG nova.network.neutron [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Successfully updated port: 1e67e320-f1e0-40f7-b0df-9cac23dbf102 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1073.168644] env[62600]: DEBUG nova.compute.manager [req-3cfb958d-fa57-40e2-baa9-fbfb319f898e req-13fe8582-f6e6-45e6-bf49-2cda958cb538 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Received event network-vif-plugged-1e67e320-f1e0-40f7-b0df-9cac23dbf102 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1073.168889] env[62600]: DEBUG oslo_concurrency.lockutils [req-3cfb958d-fa57-40e2-baa9-fbfb319f898e req-13fe8582-f6e6-45e6-bf49-2cda958cb538 service nova] Acquiring lock "eb3c5624-0599-4e81-b056-bedf34329646-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.169131] env[62600]: DEBUG oslo_concurrency.lockutils [req-3cfb958d-fa57-40e2-baa9-fbfb319f898e req-13fe8582-f6e6-45e6-bf49-2cda958cb538 service nova] Lock "eb3c5624-0599-4e81-b056-bedf34329646-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.169317] env[62600]: DEBUG oslo_concurrency.lockutils [req-3cfb958d-fa57-40e2-baa9-fbfb319f898e req-13fe8582-f6e6-45e6-bf49-2cda958cb538 service nova] Lock "eb3c5624-0599-4e81-b056-bedf34329646-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.169502] env[62600]: DEBUG nova.compute.manager [req-3cfb958d-fa57-40e2-baa9-fbfb319f898e req-13fe8582-f6e6-45e6-bf49-2cda958cb538 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] No waiting events found dispatching network-vif-plugged-1e67e320-f1e0-40f7-b0df-9cac23dbf102 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1073.169670] env[62600]: WARNING nova.compute.manager [req-3cfb958d-fa57-40e2-baa9-fbfb319f898e req-13fe8582-f6e6-45e6-bf49-2cda958cb538 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Received unexpected event network-vif-plugged-1e67e320-f1e0-40f7-b0df-9cac23dbf102 for instance with vm_state active and task_state None. [ 1073.466351] env[62600]: DEBUG oslo_concurrency.lockutils [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.466443] env[62600]: DEBUG oslo_concurrency.lockutils [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.467630] env[62600]: DEBUG nova.network.neutron [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1073.984017] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "1d6bc00e-b18b-4333-b500-bb9d9d00d4e8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.984327] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "1d6bc00e-b18b-4333-b500-bb9d9d00d4e8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.984543] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "1d6bc00e-b18b-4333-b500-bb9d9d00d4e8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.984735] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "1d6bc00e-b18b-4333-b500-bb9d9d00d4e8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.984914] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "1d6bc00e-b18b-4333-b500-bb9d9d00d4e8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.986923] env[62600]: INFO nova.compute.manager [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Terminating instance [ 1073.988655] env[62600]: DEBUG nova.compute.manager [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1073.988858] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1073.989681] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-320fffca-82a2-44af-8462-887248d5cc0d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.997533] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1073.997768] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4f028c0b-11f2-4023-98cd-790b952fad90 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.000683] env[62600]: WARNING nova.network.neutron [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] 8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb already exists in list: networks containing: ['8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb']. ignoring it [ 1074.056406] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1074.057058] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1074.057058] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleting the datastore file [datastore2] 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1074.057193] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-39431df8-f034-498b-ae95-6d7d87487c74 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.063931] env[62600]: DEBUG oslo_vmware.api [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1074.063931] env[62600]: value = "task-1223032" [ 1074.063931] env[62600]: _type = "Task" [ 1074.063931] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.071656] env[62600]: DEBUG oslo_vmware.api [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223032, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.271253] env[62600]: DEBUG nova.network.neutron [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Updating instance_info_cache with network_info: [{"id": "b8e9cd2e-1990-4997-b1fa-cb4411d39fde", "address": "fa:16:3e:fc:c1:ac", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8e9cd2e-19", "ovs_interfaceid": "b8e9cd2e-1990-4997-b1fa-cb4411d39fde", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1e67e320-f1e0-40f7-b0df-9cac23dbf102", "address": "fa:16:3e:d4:62:b8", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e67e320-f1", "ovs_interfaceid": "1e67e320-f1e0-40f7-b0df-9cac23dbf102", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.576161] env[62600]: DEBUG oslo_vmware.api [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223032, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.774556] env[62600]: DEBUG oslo_concurrency.lockutils [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.775379] env[62600]: DEBUG oslo_concurrency.lockutils [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.775561] env[62600]: DEBUG oslo_concurrency.lockutils [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.776491] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa6af4d0-5fb0-4e4c-8e77-34c89bad82d3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.793722] env[62600]: DEBUG nova.virt.hardware [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1074.794011] env[62600]: DEBUG nova.virt.hardware [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1074.794197] env[62600]: DEBUG nova.virt.hardware [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1074.794411] env[62600]: DEBUG nova.virt.hardware [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1074.794569] env[62600]: DEBUG nova.virt.hardware [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1074.794756] env[62600]: DEBUG nova.virt.hardware [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1074.795042] env[62600]: DEBUG nova.virt.hardware [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1074.795231] env[62600]: DEBUG nova.virt.hardware [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1074.795411] env[62600]: DEBUG nova.virt.hardware [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1074.795581] env[62600]: DEBUG nova.virt.hardware [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1074.795759] env[62600]: DEBUG nova.virt.hardware [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1074.801931] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Reconfiguring VM to attach interface {{(pid=62600) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1074.802285] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8adfc0e8-8b1f-48d6-a740-a5daa2c50493 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.820431] env[62600]: DEBUG oslo_vmware.api [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1074.820431] env[62600]: value = "task-1223033" [ 1074.820431] env[62600]: _type = "Task" [ 1074.820431] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.827878] env[62600]: DEBUG oslo_vmware.api [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223033, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.076103] env[62600]: DEBUG oslo_vmware.api [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223032, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.619341} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.076407] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1075.076606] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1075.076795] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1075.076983] env[62600]: INFO nova.compute.manager [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Took 1.09 seconds to destroy the instance on the hypervisor. [ 1075.077389] env[62600]: DEBUG oslo.service.loopingcall [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1075.077460] env[62600]: DEBUG nova.compute.manager [-] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1075.077542] env[62600]: DEBUG nova.network.neutron [-] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1075.201497] env[62600]: DEBUG nova.compute.manager [req-4b3efb64-6f21-45dc-88de-8e0d545a84ff req-8b3bd5f1-9280-448f-82b6-cfa02bc3a327 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Received event network-changed-1e67e320-f1e0-40f7-b0df-9cac23dbf102 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1075.201702] env[62600]: DEBUG nova.compute.manager [req-4b3efb64-6f21-45dc-88de-8e0d545a84ff req-8b3bd5f1-9280-448f-82b6-cfa02bc3a327 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Refreshing instance network info cache due to event network-changed-1e67e320-f1e0-40f7-b0df-9cac23dbf102. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1075.201922] env[62600]: DEBUG oslo_concurrency.lockutils [req-4b3efb64-6f21-45dc-88de-8e0d545a84ff req-8b3bd5f1-9280-448f-82b6-cfa02bc3a327 service nova] Acquiring lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.202571] env[62600]: DEBUG oslo_concurrency.lockutils [req-4b3efb64-6f21-45dc-88de-8e0d545a84ff req-8b3bd5f1-9280-448f-82b6-cfa02bc3a327 service nova] Acquired lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.203264] env[62600]: DEBUG nova.network.neutron [req-4b3efb64-6f21-45dc-88de-8e0d545a84ff req-8b3bd5f1-9280-448f-82b6-cfa02bc3a327 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Refreshing network info cache for port 1e67e320-f1e0-40f7-b0df-9cac23dbf102 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1075.331136] env[62600]: DEBUG oslo_vmware.api [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223033, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.786058] env[62600]: DEBUG nova.network.neutron [-] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.831240] env[62600]: DEBUG oslo_vmware.api [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223033, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.913448] env[62600]: DEBUG nova.network.neutron [req-4b3efb64-6f21-45dc-88de-8e0d545a84ff req-8b3bd5f1-9280-448f-82b6-cfa02bc3a327 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Updated VIF entry in instance network info cache for port 1e67e320-f1e0-40f7-b0df-9cac23dbf102. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1075.913884] env[62600]: DEBUG nova.network.neutron [req-4b3efb64-6f21-45dc-88de-8e0d545a84ff req-8b3bd5f1-9280-448f-82b6-cfa02bc3a327 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Updating instance_info_cache with network_info: [{"id": "b8e9cd2e-1990-4997-b1fa-cb4411d39fde", "address": "fa:16:3e:fc:c1:ac", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8e9cd2e-19", "ovs_interfaceid": "b8e9cd2e-1990-4997-b1fa-cb4411d39fde", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1e67e320-f1e0-40f7-b0df-9cac23dbf102", "address": "fa:16:3e:d4:62:b8", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e67e320-f1", "ovs_interfaceid": "1e67e320-f1e0-40f7-b0df-9cac23dbf102", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.287890] env[62600]: INFO nova.compute.manager [-] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Took 1.21 seconds to deallocate network for instance. [ 1076.332119] env[62600]: DEBUG oslo_vmware.api [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223033, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.416968] env[62600]: DEBUG oslo_concurrency.lockutils [req-4b3efb64-6f21-45dc-88de-8e0d545a84ff req-8b3bd5f1-9280-448f-82b6-cfa02bc3a327 service nova] Releasing lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.794309] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.794620] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.794855] env[62600]: DEBUG nova.objects.instance [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lazy-loading 'resources' on Instance uuid 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1076.832164] env[62600]: DEBUG oslo_vmware.api [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223033, 'name': ReconfigVM_Task, 'duration_secs': 1.616833} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.832643] env[62600]: DEBUG oslo_concurrency.lockutils [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.832857] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Reconfigured VM to attach interface {{(pid=62600) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1077.233800] env[62600]: DEBUG nova.compute.manager [req-9678bef8-1149-44d9-bf98-c0f2054489b3 req-c676cdff-e320-4796-a444-4b90d3717a49 service nova] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Received event network-vif-deleted-c39361b1-8fba-41ad-a28f-a72eeafde1e2 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1077.337614] env[62600]: DEBUG oslo_concurrency.lockutils [None req-581ba21c-b7fa-4a45-bb8a-8cd6832fd9df tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "interface-eb3c5624-0599-4e81-b056-bedf34329646-1e67e320-f1e0-40f7-b0df-9cac23dbf102" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.999s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.392571] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6427f569-0108-434e-966e-0de779db3c5e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.400513] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d59ac10b-503f-41d2-aa0e-23344fe0b373 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.430681] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e45b54-2263-4c43-8cb0-cf5639bc2308 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.437957] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195fd1be-bdb6-49b8-b59d-b04a4f635cce {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.450839] env[62600]: DEBUG nova.compute.provider_tree [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1077.954285] env[62600]: DEBUG nova.scheduler.client.report [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1078.458907] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.664s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.479422] env[62600]: INFO nova.scheduler.client.report [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleted allocations for instance 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8 [ 1078.708511] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "interface-eb3c5624-0599-4e81-b056-bedf34329646-1e67e320-f1e0-40f7-b0df-9cac23dbf102" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.708805] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "interface-eb3c5624-0599-4e81-b056-bedf34329646-1e67e320-f1e0-40f7-b0df-9cac23dbf102" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.986394] env[62600]: DEBUG oslo_concurrency.lockutils [None req-0153432d-b0ea-4dd0-b79b-3113e68b4ccc tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "1d6bc00e-b18b-4333-b500-bb9d9d00d4e8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.002s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.211786] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.211998] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.212920] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7c654c4-425c-49fa-9b5a-de6a1eebc852 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.230354] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c782c2e-5449-44d6-b937-750674cbd423 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.255514] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Reconfiguring VM to detach interface {{(pid=62600) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1079.255921] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e519e8cd-a53d-4a94-9096-b403b33d3310 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.278853] env[62600]: DEBUG oslo_vmware.api [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1079.278853] env[62600]: value = "task-1223034" [ 1079.278853] env[62600]: _type = "Task" [ 1079.278853] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.287054] env[62600]: DEBUG oslo_vmware.api [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223034, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.789415] env[62600]: DEBUG oslo_vmware.api [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223034, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.019491] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "1b856aff-66be-4fa5-b8b7-124ac24fb4d1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.019761] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "1b856aff-66be-4fa5-b8b7-124ac24fb4d1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.019980] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "1b856aff-66be-4fa5-b8b7-124ac24fb4d1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.020191] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "1b856aff-66be-4fa5-b8b7-124ac24fb4d1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.020368] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "1b856aff-66be-4fa5-b8b7-124ac24fb4d1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.022522] env[62600]: INFO nova.compute.manager [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Terminating instance [ 1080.024411] env[62600]: DEBUG nova.compute.manager [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1080.024619] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1080.025631] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc388ac-1c11-4232-8165-e5e7e041f3ef {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.035014] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1080.035255] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-304a508a-c9ca-461a-bf0f-edf254b9f1ab {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.041642] env[62600]: DEBUG oslo_vmware.api [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1080.041642] env[62600]: value = "task-1223035" [ 1080.041642] env[62600]: _type = "Task" [ 1080.041642] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.048984] env[62600]: DEBUG oslo_vmware.api [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223035, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.289111] env[62600]: DEBUG oslo_vmware.api [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223034, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.552068] env[62600]: DEBUG oslo_vmware.api [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223035, 'name': PowerOffVM_Task, 'duration_secs': 0.193849} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.552319] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1080.552494] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1080.552752] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ac2c7acc-ba35-4bb7-b43a-d063bb60d4f3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.610871] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1080.611150] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1080.611419] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleting the datastore file [datastore2] 1b856aff-66be-4fa5-b8b7-124ac24fb4d1 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1080.611721] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-15995b9e-a607-44d9-8adb-401cb2823397 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.617429] env[62600]: DEBUG oslo_vmware.api [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for the task: (returnval){ [ 1080.617429] env[62600]: value = "task-1223037" [ 1080.617429] env[62600]: _type = "Task" [ 1080.617429] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.624795] env[62600]: DEBUG oslo_vmware.api [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223037, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.746558] env[62600]: DEBUG oslo_concurrency.lockutils [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "8417575c-d68d-464f-8c22-ab358deb76d8" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.746832] env[62600]: DEBUG oslo_concurrency.lockutils [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "8417575c-d68d-464f-8c22-ab358deb76d8" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.789755] env[62600]: DEBUG oslo_vmware.api [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223034, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.127843] env[62600]: DEBUG oslo_vmware.api [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Task: {'id': task-1223037, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127502} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.128093] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1081.128357] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1081.128555] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1081.128741] env[62600]: INFO nova.compute.manager [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1081.129009] env[62600]: DEBUG oslo.service.loopingcall [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1081.129230] env[62600]: DEBUG nova.compute.manager [-] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1081.129325] env[62600]: DEBUG nova.network.neutron [-] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1081.250211] env[62600]: DEBUG nova.compute.utils [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1081.289576] env[62600]: DEBUG oslo_vmware.api [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223034, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.352468] env[62600]: DEBUG nova.compute.manager [req-7eb4baa3-7edd-46da-8a12-95a41eb7ad83 req-d89f61d9-e7f8-42a5-a70d-979668af6b40 service nova] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Received event network-vif-deleted-30cc50bd-a129-4992-8ced-66359d34fcc1 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1081.352468] env[62600]: INFO nova.compute.manager [req-7eb4baa3-7edd-46da-8a12-95a41eb7ad83 req-d89f61d9-e7f8-42a5-a70d-979668af6b40 service nova] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Neutron deleted interface 30cc50bd-a129-4992-8ced-66359d34fcc1; detaching it from the instance and deleting it from the info cache [ 1081.352570] env[62600]: DEBUG nova.network.neutron [req-7eb4baa3-7edd-46da-8a12-95a41eb7ad83 req-d89f61d9-e7f8-42a5-a70d-979668af6b40 service nova] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.752976] env[62600]: DEBUG oslo_concurrency.lockutils [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "8417575c-d68d-464f-8c22-ab358deb76d8" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.790883] env[62600]: DEBUG oslo_vmware.api [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223034, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.831117] env[62600]: DEBUG nova.network.neutron [-] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.855039] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8d777620-a22f-4ade-9545-bfc1d03d2247 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.864338] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2de05b6c-572a-481d-85f8-e0b969e54bad {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.890277] env[62600]: DEBUG nova.compute.manager [req-7eb4baa3-7edd-46da-8a12-95a41eb7ad83 req-d89f61d9-e7f8-42a5-a70d-979668af6b40 service nova] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Detach interface failed, port_id=30cc50bd-a129-4992-8ced-66359d34fcc1, reason: Instance 1b856aff-66be-4fa5-b8b7-124ac24fb4d1 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1082.291240] env[62600]: DEBUG oslo_vmware.api [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223034, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.335018] env[62600]: INFO nova.compute.manager [-] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Took 1.21 seconds to deallocate network for instance. [ 1082.791942] env[62600]: DEBUG oslo_vmware.api [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223034, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.814737] env[62600]: DEBUG oslo_concurrency.lockutils [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "8417575c-d68d-464f-8c22-ab358deb76d8" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.814990] env[62600]: DEBUG oslo_concurrency.lockutils [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "8417575c-d68d-464f-8c22-ab358deb76d8" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.815269] env[62600]: INFO nova.compute.manager [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Attaching volume 3e4dff37-0176-459b-a9f5-7b58fc036685 to /dev/sdb [ 1082.841464] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.841726] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.841949] env[62600]: DEBUG nova.objects.instance [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lazy-loading 'resources' on Instance uuid 1b856aff-66be-4fa5-b8b7-124ac24fb4d1 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1082.844933] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1834f56e-6adb-4c03-8b83-589e8faf05d2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.851658] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fee79b3-1047-456c-8675-c8c4585041bd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.864240] env[62600]: DEBUG nova.virt.block_device [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Updating existing volume attachment record: de9db270-40cc-40a8-9222-e763e0a3d0c5 {{(pid=62600) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1083.293301] env[62600]: DEBUG oslo_vmware.api [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223034, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.429533] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc6bd2f0-b5fa-4c75-8c07-81d3fce2ded4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.436999] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57c9afd8-75db-4983-8fb0-194b993ac826 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.467513] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c6340c-d36f-4bee-9a32-db4ffcca3702 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.474069] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb24a519-6ad9-48b2-8f2b-a264066d36b8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.486463] env[62600]: DEBUG nova.compute.provider_tree [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1083.793222] env[62600]: DEBUG oslo_vmware.api [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223034, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.989679] env[62600]: DEBUG nova.scheduler.client.report [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1084.293924] env[62600]: DEBUG oslo_vmware.api [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223034, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.494695] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.653s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.514373] env[62600]: INFO nova.scheduler.client.report [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Deleted allocations for instance 1b856aff-66be-4fa5-b8b7-124ac24fb4d1 [ 1084.794290] env[62600]: DEBUG oslo_vmware.api [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223034, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.022215] env[62600]: DEBUG oslo_concurrency.lockutils [None req-7e7274e5-bc21-462f-9342-298f8c214d31 tempest-ServersTestJSON-736885129 tempest-ServersTestJSON-736885129-project-member] Lock "1b856aff-66be-4fa5-b8b7-124ac24fb4d1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.002s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.295554] env[62600]: DEBUG oslo_vmware.api [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223034, 'name': ReconfigVM_Task, 'duration_secs': 5.73746} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.295813] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.296072] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Reconfigured VM to detach interface {{(pid=62600) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1086.616788] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.617050] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.617210] env[62600]: DEBUG nova.network.neutron [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1087.411554] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Volume attach. Driver type: vmdk {{(pid=62600) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1087.411797] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264372', 'volume_id': '3e4dff37-0176-459b-a9f5-7b58fc036685', 'name': 'volume-3e4dff37-0176-459b-a9f5-7b58fc036685', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8417575c-d68d-464f-8c22-ab358deb76d8', 'attached_at': '', 'detached_at': '', 'volume_id': '3e4dff37-0176-459b-a9f5-7b58fc036685', 'serial': '3e4dff37-0176-459b-a9f5-7b58fc036685'} {{(pid=62600) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1087.412705] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e28991a-5763-4cf9-9a03-40c3eef27b59 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.429133] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-092e48b2-1219-4b76-99d1-931d3c242fa0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.453445] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] volume-3e4dff37-0176-459b-a9f5-7b58fc036685/volume-3e4dff37-0176-459b-a9f5-7b58fc036685.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1087.453728] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a11cf5f9-470a-44d5-8bdc-a6eb3bbad8b3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.467381] env[62600]: INFO nova.network.neutron [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Port 1e67e320-f1e0-40f7-b0df-9cac23dbf102 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1087.467717] env[62600]: DEBUG nova.network.neutron [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Updating instance_info_cache with network_info: [{"id": "b8e9cd2e-1990-4997-b1fa-cb4411d39fde", "address": "fa:16:3e:fc:c1:ac", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8e9cd2e-19", "ovs_interfaceid": "b8e9cd2e-1990-4997-b1fa-cb4411d39fde", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.474715] env[62600]: DEBUG oslo_vmware.api [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1087.474715] env[62600]: value = "task-1223040" [ 1087.474715] env[62600]: _type = "Task" [ 1087.474715] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.482693] env[62600]: DEBUG oslo_vmware.api [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223040, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.505119] env[62600]: DEBUG nova.compute.manager [req-208aa94f-4d97-4bc6-afc0-bcd33994de8e req-8e16d7b4-673d-4485-add1-17d76d254bf2 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Received event network-changed-b8e9cd2e-1990-4997-b1fa-cb4411d39fde {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1087.505376] env[62600]: DEBUG nova.compute.manager [req-208aa94f-4d97-4bc6-afc0-bcd33994de8e req-8e16d7b4-673d-4485-add1-17d76d254bf2 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Refreshing instance network info cache due to event network-changed-b8e9cd2e-1990-4997-b1fa-cb4411d39fde. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1087.505947] env[62600]: DEBUG oslo_concurrency.lockutils [req-208aa94f-4d97-4bc6-afc0-bcd33994de8e req-8e16d7b4-673d-4485-add1-17d76d254bf2 service nova] Acquiring lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.970486] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.972641] env[62600]: DEBUG oslo_concurrency.lockutils [req-208aa94f-4d97-4bc6-afc0-bcd33994de8e req-8e16d7b4-673d-4485-add1-17d76d254bf2 service nova] Acquired lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.972851] env[62600]: DEBUG nova.network.neutron [req-208aa94f-4d97-4bc6-afc0-bcd33994de8e req-8e16d7b4-673d-4485-add1-17d76d254bf2 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Refreshing network info cache for port b8e9cd2e-1990-4997-b1fa-cb4411d39fde {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1087.985743] env[62600]: DEBUG oslo_vmware.api [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223040, 'name': ReconfigVM_Task, 'duration_secs': 0.340161} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.986031] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Reconfigured VM instance instance-00000062 to attach disk [datastore2] volume-3e4dff37-0176-459b-a9f5-7b58fc036685/volume-3e4dff37-0176-459b-a9f5-7b58fc036685.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1087.990765] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a4399bb-01fa-4e6f-8662-0daec2df412d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.006365] env[62600]: DEBUG oslo_vmware.api [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1088.006365] env[62600]: value = "task-1223041" [ 1088.006365] env[62600]: _type = "Task" [ 1088.006365] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.013937] env[62600]: DEBUG oslo_vmware.api [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223041, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.178475] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "interface-1817183b-44a5-41ec-a2bf-818edee264b3-1e67e320-f1e0-40f7-b0df-9cac23dbf102" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.178754] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "interface-1817183b-44a5-41ec-a2bf-818edee264b3-1e67e320-f1e0-40f7-b0df-9cac23dbf102" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.179157] env[62600]: DEBUG nova.objects.instance [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lazy-loading 'flavor' on Instance uuid 1817183b-44a5-41ec-a2bf-818edee264b3 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1088.475341] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d05e4f3a-7511-47d6-993e-586bc4e92ff7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "interface-eb3c5624-0599-4e81-b056-bedf34329646-1e67e320-f1e0-40f7-b0df-9cac23dbf102" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.766s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.518857] env[62600]: DEBUG oslo_vmware.api [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223041, 'name': ReconfigVM_Task, 'duration_secs': 0.155723} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.521663] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264372', 'volume_id': '3e4dff37-0176-459b-a9f5-7b58fc036685', 'name': 'volume-3e4dff37-0176-459b-a9f5-7b58fc036685', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8417575c-d68d-464f-8c22-ab358deb76d8', 'attached_at': '', 'detached_at': '', 'volume_id': '3e4dff37-0176-459b-a9f5-7b58fc036685', 'serial': '3e4dff37-0176-459b-a9f5-7b58fc036685'} {{(pid=62600) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1088.688820] env[62600]: DEBUG nova.network.neutron [req-208aa94f-4d97-4bc6-afc0-bcd33994de8e req-8e16d7b4-673d-4485-add1-17d76d254bf2 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Updated VIF entry in instance network info cache for port b8e9cd2e-1990-4997-b1fa-cb4411d39fde. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1088.689305] env[62600]: DEBUG nova.network.neutron [req-208aa94f-4d97-4bc6-afc0-bcd33994de8e req-8e16d7b4-673d-4485-add1-17d76d254bf2 service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Updating instance_info_cache with network_info: [{"id": "b8e9cd2e-1990-4997-b1fa-cb4411d39fde", "address": "fa:16:3e:fc:c1:ac", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8e9cd2e-19", "ovs_interfaceid": "b8e9cd2e-1990-4997-b1fa-cb4411d39fde", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.772319] env[62600]: DEBUG nova.objects.instance [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lazy-loading 'pci_requests' on Instance uuid 1817183b-44a5-41ec-a2bf-818edee264b3 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.192310] env[62600]: DEBUG oslo_concurrency.lockutils [req-208aa94f-4d97-4bc6-afc0-bcd33994de8e req-8e16d7b4-673d-4485-add1-17d76d254bf2 service nova] Releasing lock "refresh_cache-eb3c5624-0599-4e81-b056-bedf34329646" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.275161] env[62600]: DEBUG nova.objects.base [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Object Instance<1817183b-44a5-41ec-a2bf-818edee264b3> lazy-loaded attributes: flavor,pci_requests {{(pid=62600) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1089.275401] env[62600]: DEBUG nova.network.neutron [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1089.344646] env[62600]: DEBUG nova.policy [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fbd1b2fb34d841359ada8fc44bec2986', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b028450f2da445fb83e37adfc86bba68', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 1089.535444] env[62600]: DEBUG nova.compute.manager [req-ea518ef5-480f-45b6-82c6-a51a49596f8d req-96507f8e-66c6-42ad-8cfc-7e1ef2e1e4d2 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Received event network-changed-5796bb3c-0926-47b4-bcdb-c54c18f83200 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1089.535444] env[62600]: DEBUG nova.compute.manager [req-ea518ef5-480f-45b6-82c6-a51a49596f8d req-96507f8e-66c6-42ad-8cfc-7e1ef2e1e4d2 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Refreshing instance network info cache due to event network-changed-5796bb3c-0926-47b4-bcdb-c54c18f83200. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1089.535444] env[62600]: DEBUG oslo_concurrency.lockutils [req-ea518ef5-480f-45b6-82c6-a51a49596f8d req-96507f8e-66c6-42ad-8cfc-7e1ef2e1e4d2 service nova] Acquiring lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.535444] env[62600]: DEBUG oslo_concurrency.lockutils [req-ea518ef5-480f-45b6-82c6-a51a49596f8d req-96507f8e-66c6-42ad-8cfc-7e1ef2e1e4d2 service nova] Acquired lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.535444] env[62600]: DEBUG nova.network.neutron [req-ea518ef5-480f-45b6-82c6-a51a49596f8d req-96507f8e-66c6-42ad-8cfc-7e1ef2e1e4d2 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Refreshing network info cache for port 5796bb3c-0926-47b4-bcdb-c54c18f83200 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1089.556021] env[62600]: DEBUG nova.objects.instance [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lazy-loading 'flavor' on Instance uuid 8417575c-d68d-464f-8c22-ab358deb76d8 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1090.060472] env[62600]: DEBUG oslo_concurrency.lockutils [None req-426ab503-84c1-494b-a503-ecb8d02e1b98 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "8417575c-d68d-464f-8c22-ab358deb76d8" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.245s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.259736] env[62600]: DEBUG nova.network.neutron [req-ea518ef5-480f-45b6-82c6-a51a49596f8d req-96507f8e-66c6-42ad-8cfc-7e1ef2e1e4d2 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Updated VIF entry in instance network info cache for port 5796bb3c-0926-47b4-bcdb-c54c18f83200. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1090.260198] env[62600]: DEBUG nova.network.neutron [req-ea518ef5-480f-45b6-82c6-a51a49596f8d req-96507f8e-66c6-42ad-8cfc-7e1ef2e1e4d2 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Updating instance_info_cache with network_info: [{"id": "5796bb3c-0926-47b4-bcdb-c54c18f83200", "address": "fa:16:3e:94:99:6b", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5796bb3c-09", "ovs_interfaceid": "5796bb3c-0926-47b4-bcdb-c54c18f83200", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.314213] env[62600]: DEBUG oslo_concurrency.lockutils [None req-02a55929-6b64-470c-a76f-a5b329117622 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "8417575c-d68d-464f-8c22-ab358deb76d8" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.314493] env[62600]: DEBUG oslo_concurrency.lockutils [None req-02a55929-6b64-470c-a76f-a5b329117622 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "8417575c-d68d-464f-8c22-ab358deb76d8" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.735068] env[62600]: DEBUG nova.compute.manager [req-4dd5750c-c920-4aeb-9b84-80f1457e7e57 req-c4a1f4fb-df98-4cf8-9638-4296a61df269 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Received event network-vif-plugged-1e67e320-f1e0-40f7-b0df-9cac23dbf102 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1090.735231] env[62600]: DEBUG oslo_concurrency.lockutils [req-4dd5750c-c920-4aeb-9b84-80f1457e7e57 req-c4a1f4fb-df98-4cf8-9638-4296a61df269 service nova] Acquiring lock "1817183b-44a5-41ec-a2bf-818edee264b3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.735428] env[62600]: DEBUG oslo_concurrency.lockutils [req-4dd5750c-c920-4aeb-9b84-80f1457e7e57 req-c4a1f4fb-df98-4cf8-9638-4296a61df269 service nova] Lock "1817183b-44a5-41ec-a2bf-818edee264b3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.735606] env[62600]: DEBUG oslo_concurrency.lockutils [req-4dd5750c-c920-4aeb-9b84-80f1457e7e57 req-c4a1f4fb-df98-4cf8-9638-4296a61df269 service nova] Lock "1817183b-44a5-41ec-a2bf-818edee264b3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.735782] env[62600]: DEBUG nova.compute.manager [req-4dd5750c-c920-4aeb-9b84-80f1457e7e57 req-c4a1f4fb-df98-4cf8-9638-4296a61df269 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] No waiting events found dispatching network-vif-plugged-1e67e320-f1e0-40f7-b0df-9cac23dbf102 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1090.735967] env[62600]: WARNING nova.compute.manager [req-4dd5750c-c920-4aeb-9b84-80f1457e7e57 req-c4a1f4fb-df98-4cf8-9638-4296a61df269 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Received unexpected event network-vif-plugged-1e67e320-f1e0-40f7-b0df-9cac23dbf102 for instance with vm_state active and task_state None. [ 1090.762955] env[62600]: DEBUG oslo_concurrency.lockutils [req-ea518ef5-480f-45b6-82c6-a51a49596f8d req-96507f8e-66c6-42ad-8cfc-7e1ef2e1e4d2 service nova] Releasing lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1090.817931] env[62600]: INFO nova.compute.manager [None req-02a55929-6b64-470c-a76f-a5b329117622 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Detaching volume 3e4dff37-0176-459b-a9f5-7b58fc036685 [ 1090.820667] env[62600]: DEBUG nova.network.neutron [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Successfully updated port: 1e67e320-f1e0-40f7-b0df-9cac23dbf102 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1090.853860] env[62600]: INFO nova.virt.block_device [None req-02a55929-6b64-470c-a76f-a5b329117622 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Attempting to driver detach volume 3e4dff37-0176-459b-a9f5-7b58fc036685 from mountpoint /dev/sdb [ 1090.854233] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-02a55929-6b64-470c-a76f-a5b329117622 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Volume detach. Driver type: vmdk {{(pid=62600) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1090.854346] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-02a55929-6b64-470c-a76f-a5b329117622 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264372', 'volume_id': '3e4dff37-0176-459b-a9f5-7b58fc036685', 'name': 'volume-3e4dff37-0176-459b-a9f5-7b58fc036685', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8417575c-d68d-464f-8c22-ab358deb76d8', 'attached_at': '', 'detached_at': '', 'volume_id': '3e4dff37-0176-459b-a9f5-7b58fc036685', 'serial': '3e4dff37-0176-459b-a9f5-7b58fc036685'} {{(pid=62600) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1090.855265] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b36558e-8290-4ec9-9dc8-3fd5a85ba982 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.878563] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff75809-3393-491c-a5d9-774ba1806a55 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.885143] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32c4eef0-1e70-46e7-8895-c4c4bc7d3136 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.904568] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b88f7a-cebb-404d-80e5-06ad8c0d2845 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.918465] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-02a55929-6b64-470c-a76f-a5b329117622 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] The volume has not been displaced from its original location: [datastore2] volume-3e4dff37-0176-459b-a9f5-7b58fc036685/volume-3e4dff37-0176-459b-a9f5-7b58fc036685.vmdk. No consolidation needed. {{(pid=62600) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1090.923682] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-02a55929-6b64-470c-a76f-a5b329117622 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Reconfiguring VM instance instance-00000062 to detach disk 2001 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1090.923932] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c03c25f-e96e-4f20-adf0-741d60cb2a6a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.941052] env[62600]: DEBUG oslo_vmware.api [None req-02a55929-6b64-470c-a76f-a5b329117622 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1090.941052] env[62600]: value = "task-1223042" [ 1090.941052] env[62600]: _type = "Task" [ 1090.941052] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.948275] env[62600]: DEBUG oslo_vmware.api [None req-02a55929-6b64-470c-a76f-a5b329117622 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223042, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.323793] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.324239] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.324239] env[62600]: DEBUG nova.network.neutron [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1091.450734] env[62600]: DEBUG oslo_vmware.api [None req-02a55929-6b64-470c-a76f-a5b329117622 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223042, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.951903] env[62600]: DEBUG oslo_vmware.api [None req-02a55929-6b64-470c-a76f-a5b329117622 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223042, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.032898] env[62600]: WARNING nova.network.neutron [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] 8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb already exists in list: networks containing: ['8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb']. ignoring it [ 1092.294392] env[62600]: DEBUG nova.network.neutron [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Updating instance_info_cache with network_info: [{"id": "5796bb3c-0926-47b4-bcdb-c54c18f83200", "address": "fa:16:3e:94:99:6b", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5796bb3c-09", "ovs_interfaceid": "5796bb3c-0926-47b4-bcdb-c54c18f83200", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1e67e320-f1e0-40f7-b0df-9cac23dbf102", "address": "fa:16:3e:d4:62:b8", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e67e320-f1", "ovs_interfaceid": "1e67e320-f1e0-40f7-b0df-9cac23dbf102", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.429378] env[62600]: DEBUG oslo_concurrency.lockutils [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "692fa608-39f4-4d7d-9f30-b6559878b970" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.429706] env[62600]: DEBUG oslo_concurrency.lockutils [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "692fa608-39f4-4d7d-9f30-b6559878b970" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.429821] env[62600]: INFO nova.compute.manager [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Shelving [ 1092.451846] env[62600]: DEBUG oslo_vmware.api [None req-02a55929-6b64-470c-a76f-a5b329117622 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223042, 'name': ReconfigVM_Task, 'duration_secs': 1.213424} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.452142] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-02a55929-6b64-470c-a76f-a5b329117622 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Reconfigured VM instance instance-00000062 to detach disk 2001 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1092.456841] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71e7f2c0-ad04-4f27-af41-848bd75f2756 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.470818] env[62600]: DEBUG oslo_vmware.api [None req-02a55929-6b64-470c-a76f-a5b329117622 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1092.470818] env[62600]: value = "task-1223043" [ 1092.470818] env[62600]: _type = "Task" [ 1092.470818] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.478583] env[62600]: DEBUG oslo_vmware.api [None req-02a55929-6b64-470c-a76f-a5b329117622 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223043, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.763292] env[62600]: DEBUG nova.compute.manager [req-c2ace7e8-b0be-4e38-9cc7-e9d8ce6fdce6 req-8fd7caa4-c7e4-4bc1-bf09-9f22aa8410b3 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Received event network-changed-1e67e320-f1e0-40f7-b0df-9cac23dbf102 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1092.763533] env[62600]: DEBUG nova.compute.manager [req-c2ace7e8-b0be-4e38-9cc7-e9d8ce6fdce6 req-8fd7caa4-c7e4-4bc1-bf09-9f22aa8410b3 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Refreshing instance network info cache due to event network-changed-1e67e320-f1e0-40f7-b0df-9cac23dbf102. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1092.763759] env[62600]: DEBUG oslo_concurrency.lockutils [req-c2ace7e8-b0be-4e38-9cc7-e9d8ce6fdce6 req-8fd7caa4-c7e4-4bc1-bf09-9f22aa8410b3 service nova] Acquiring lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.797425] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.798296] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.798296] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.798491] env[62600]: DEBUG oslo_concurrency.lockutils [req-c2ace7e8-b0be-4e38-9cc7-e9d8ce6fdce6 req-8fd7caa4-c7e4-4bc1-bf09-9f22aa8410b3 service nova] Acquired lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.798670] env[62600]: DEBUG nova.network.neutron [req-c2ace7e8-b0be-4e38-9cc7-e9d8ce6fdce6 req-8fd7caa4-c7e4-4bc1-bf09-9f22aa8410b3 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Refreshing network info cache for port 1e67e320-f1e0-40f7-b0df-9cac23dbf102 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1092.800759] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecaad466-3ca5-4c17-beb4-8773589cb985 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.818212] env[62600]: DEBUG nova.virt.hardware [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1092.818456] env[62600]: DEBUG nova.virt.hardware [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1092.818624] env[62600]: DEBUG nova.virt.hardware [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1092.818814] env[62600]: DEBUG nova.virt.hardware [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1092.819016] env[62600]: DEBUG nova.virt.hardware [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1092.819195] env[62600]: DEBUG nova.virt.hardware [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1092.819416] env[62600]: DEBUG nova.virt.hardware [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1092.819585] env[62600]: DEBUG nova.virt.hardware [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1092.819760] env[62600]: DEBUG nova.virt.hardware [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1092.819931] env[62600]: DEBUG nova.virt.hardware [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1092.820129] env[62600]: DEBUG nova.virt.hardware [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1092.826324] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Reconfiguring VM to attach interface {{(pid=62600) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1092.827202] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5be644d-d7a4-464c-aad7-17193ef8c5f1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.845231] env[62600]: DEBUG oslo_vmware.api [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1092.845231] env[62600]: value = "task-1223044" [ 1092.845231] env[62600]: _type = "Task" [ 1092.845231] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.852861] env[62600]: DEBUG oslo_vmware.api [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223044, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.937696] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1092.937991] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-29b3f938-9477-4e73-95f3-aa5c771e1221 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.945517] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1092.945517] env[62600]: value = "task-1223045" [ 1092.945517] env[62600]: _type = "Task" [ 1092.945517] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.952831] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223045, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.980326] env[62600]: DEBUG oslo_vmware.api [None req-02a55929-6b64-470c-a76f-a5b329117622 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223043, 'name': ReconfigVM_Task, 'duration_secs': 0.123502} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.980841] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-02a55929-6b64-470c-a76f-a5b329117622 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264372', 'volume_id': '3e4dff37-0176-459b-a9f5-7b58fc036685', 'name': 'volume-3e4dff37-0176-459b-a9f5-7b58fc036685', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8417575c-d68d-464f-8c22-ab358deb76d8', 'attached_at': '', 'detached_at': '', 'volume_id': '3e4dff37-0176-459b-a9f5-7b58fc036685', 'serial': '3e4dff37-0176-459b-a9f5-7b58fc036685'} {{(pid=62600) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1093.354750] env[62600]: DEBUG oslo_vmware.api [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223044, 'name': ReconfigVM_Task, 'duration_secs': 0.500324} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.355440] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.355693] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Reconfigured VM to attach interface {{(pid=62600) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1093.454381] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223045, 'name': PowerOffVM_Task, 'duration_secs': 0.213532} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.454730] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1093.455458] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df89a6b-fec0-4f59-aeb5-81312056366a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.477178] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dbd7d3e-bdbd-4e2b-a4f2-4c51392100cb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.514707] env[62600]: DEBUG nova.network.neutron [req-c2ace7e8-b0be-4e38-9cc7-e9d8ce6fdce6 req-8fd7caa4-c7e4-4bc1-bf09-9f22aa8410b3 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Updated VIF entry in instance network info cache for port 1e67e320-f1e0-40f7-b0df-9cac23dbf102. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1093.515197] env[62600]: DEBUG nova.network.neutron [req-c2ace7e8-b0be-4e38-9cc7-e9d8ce6fdce6 req-8fd7caa4-c7e4-4bc1-bf09-9f22aa8410b3 service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Updating instance_info_cache with network_info: [{"id": "5796bb3c-0926-47b4-bcdb-c54c18f83200", "address": "fa:16:3e:94:99:6b", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5796bb3c-09", "ovs_interfaceid": "5796bb3c-0926-47b4-bcdb-c54c18f83200", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "1e67e320-f1e0-40f7-b0df-9cac23dbf102", "address": "fa:16:3e:d4:62:b8", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e67e320-f1", "ovs_interfaceid": "1e67e320-f1e0-40f7-b0df-9cac23dbf102", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.525732] env[62600]: DEBUG nova.objects.instance [None req-02a55929-6b64-470c-a76f-a5b329117622 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lazy-loading 'flavor' on Instance uuid 8417575c-d68d-464f-8c22-ab358deb76d8 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1093.861042] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d07cbe53-f583-43b7-bc73-be32516ac4b7 tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "interface-1817183b-44a5-41ec-a2bf-818edee264b3-1e67e320-f1e0-40f7-b0df-9cac23dbf102" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 5.682s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.988160] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Creating Snapshot of the VM instance {{(pid=62600) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1093.988489] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9a85eabf-a95c-46dc-a740-c83112dc9ab1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.996524] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1093.996524] env[62600]: value = "task-1223046" [ 1093.996524] env[62600]: _type = "Task" [ 1093.996524] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.005449] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223046, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.018233] env[62600]: DEBUG oslo_concurrency.lockutils [req-c2ace7e8-b0be-4e38-9cc7-e9d8ce6fdce6 req-8fd7caa4-c7e4-4bc1-bf09-9f22aa8410b3 service nova] Releasing lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.506886] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223046, 'name': CreateSnapshot_Task, 'duration_secs': 0.368026} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.507247] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Created Snapshot of the VM instance {{(pid=62600) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1094.507875] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4118fd8e-9677-472b-96e0-216665ec99b5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.533074] env[62600]: DEBUG oslo_concurrency.lockutils [None req-02a55929-6b64-470c-a76f-a5b329117622 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "8417575c-d68d-464f-8c22-ab358deb76d8" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.218s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.025745] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Creating linked-clone VM from snapshot {{(pid=62600) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1095.025999] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-4e729dfc-4a8f-44c9-9af5-334d71f59715 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.035064] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1095.035064] env[62600]: value = "task-1223047" [ 1095.035064] env[62600]: _type = "Task" [ 1095.035064] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.042921] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223047, 'name': CloneVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.345406] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "interface-1817183b-44a5-41ec-a2bf-818edee264b3-1e67e320-f1e0-40f7-b0df-9cac23dbf102" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.345730] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "interface-1817183b-44a5-41ec-a2bf-818edee264b3-1e67e320-f1e0-40f7-b0df-9cac23dbf102" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.545229] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223047, 'name': CloneVM_Task} progress is 94%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.659494] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "8417575c-d68d-464f-8c22-ab358deb76d8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.659853] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "8417575c-d68d-464f-8c22-ab358deb76d8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.660133] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "8417575c-d68d-464f-8c22-ab358deb76d8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.660339] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "8417575c-d68d-464f-8c22-ab358deb76d8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.660518] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "8417575c-d68d-464f-8c22-ab358deb76d8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.662762] env[62600]: INFO nova.compute.manager [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Terminating instance [ 1095.664584] env[62600]: DEBUG nova.compute.manager [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1095.664783] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1095.665633] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3348c9cc-1932-4266-9a9b-b03105931fd5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.673378] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1095.673603] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-46d51307-7d3f-43f2-9283-ec53b38a5c20 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.680239] env[62600]: DEBUG oslo_vmware.api [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1095.680239] env[62600]: value = "task-1223048" [ 1095.680239] env[62600]: _type = "Task" [ 1095.680239] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.687590] env[62600]: DEBUG oslo_vmware.api [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223048, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.849437] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.849437] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.849802] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c124e22e-90cb-41a4-b4f6-3a42671d0054 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.867526] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-483aabc6-9241-4aa3-953c-2f86c4e2d3ca {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.893160] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Reconfiguring VM to detach interface {{(pid=62600) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1095.893483] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-957eaab8-4f06-4414-8392-efed3b7fefe9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.912033] env[62600]: DEBUG oslo_vmware.api [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1095.912033] env[62600]: value = "task-1223049" [ 1095.912033] env[62600]: _type = "Task" [ 1095.912033] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.919048] env[62600]: DEBUG oslo_vmware.api [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223049, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.045879] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223047, 'name': CloneVM_Task, 'duration_secs': 0.938168} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.046583] env[62600]: INFO nova.virt.vmwareapi.vmops [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Created linked-clone VM from snapshot [ 1096.047156] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d486985d-28f1-441b-aada-b4e64252c8dc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.054417] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Uploading image 10f7ff7d-b198-418a-8817-0b212a53d0a9 {{(pid=62600) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1096.077170] env[62600]: DEBUG oslo_vmware.rw_handles [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1096.077170] env[62600]: value = "vm-264374" [ 1096.077170] env[62600]: _type = "VirtualMachine" [ 1096.077170] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1096.077426] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e04f2b21-9a75-4e08-b3e4-03be1c22b6b7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.084666] env[62600]: DEBUG oslo_vmware.rw_handles [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lease: (returnval){ [ 1096.084666] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526bd145-8aa7-5f3f-e083-8e1b8cbae471" [ 1096.084666] env[62600]: _type = "HttpNfcLease" [ 1096.084666] env[62600]: } obtained for exporting VM: (result){ [ 1096.084666] env[62600]: value = "vm-264374" [ 1096.084666] env[62600]: _type = "VirtualMachine" [ 1096.084666] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1096.084914] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the lease: (returnval){ [ 1096.084914] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526bd145-8aa7-5f3f-e083-8e1b8cbae471" [ 1096.084914] env[62600]: _type = "HttpNfcLease" [ 1096.084914] env[62600]: } to be ready. {{(pid=62600) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1096.090552] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1096.090552] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526bd145-8aa7-5f3f-e083-8e1b8cbae471" [ 1096.090552] env[62600]: _type = "HttpNfcLease" [ 1096.090552] env[62600]: } is initializing. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1096.191314] env[62600]: DEBUG oslo_vmware.api [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223048, 'name': PowerOffVM_Task, 'duration_secs': 0.187767} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.191577] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1096.191749] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1096.191989] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a08c5182-b1c2-4139-bd28-7a2bee5b25fd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.275658] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1096.275965] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1096.276274] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Deleting the datastore file [datastore1] 8417575c-d68d-464f-8c22-ab358deb76d8 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1096.276617] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-75fc87a0-8b52-410d-a43e-3a1d3f21eeaf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.283781] env[62600]: DEBUG oslo_vmware.api [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1096.283781] env[62600]: value = "task-1223052" [ 1096.283781] env[62600]: _type = "Task" [ 1096.283781] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.291418] env[62600]: DEBUG oslo_vmware.api [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223052, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.422242] env[62600]: DEBUG oslo_vmware.api [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.593493] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1096.593493] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526bd145-8aa7-5f3f-e083-8e1b8cbae471" [ 1096.593493] env[62600]: _type = "HttpNfcLease" [ 1096.593493] env[62600]: } is ready. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1096.593887] env[62600]: DEBUG oslo_vmware.rw_handles [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1096.593887] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]526bd145-8aa7-5f3f-e083-8e1b8cbae471" [ 1096.593887] env[62600]: _type = "HttpNfcLease" [ 1096.593887] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1096.594595] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248e7285-7f51-46c4-ab40-540054fc9bc5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.601368] env[62600]: DEBUG oslo_vmware.rw_handles [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5232e42b-2732-3783-2af2-e77e3774d9af/disk-0.vmdk from lease info. {{(pid=62600) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1096.601589] env[62600]: DEBUG oslo_vmware.rw_handles [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5232e42b-2732-3783-2af2-e77e3774d9af/disk-0.vmdk for reading. {{(pid=62600) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1096.687831] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-72b78262-d39b-4e8e-b4c5-003bf06e9e56 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.793369] env[62600]: DEBUG oslo_vmware.api [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223052, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141413} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.793629] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1096.793822] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1096.794011] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1096.794198] env[62600]: INFO nova.compute.manager [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1096.794445] env[62600]: DEBUG oslo.service.loopingcall [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1096.794637] env[62600]: DEBUG nova.compute.manager [-] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1096.794730] env[62600]: DEBUG nova.network.neutron [-] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1096.921663] env[62600]: DEBUG oslo_vmware.api [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.266592] env[62600]: DEBUG nova.compute.manager [req-7723df6b-39d2-45b3-a58d-a70cccf0cf8f req-77d6360a-5c6e-4233-b978-cf2a45f99066 service nova] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Received event network-vif-deleted-028950a6-9715-4e67-b31b-447158a62875 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1097.266871] env[62600]: INFO nova.compute.manager [req-7723df6b-39d2-45b3-a58d-a70cccf0cf8f req-77d6360a-5c6e-4233-b978-cf2a45f99066 service nova] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Neutron deleted interface 028950a6-9715-4e67-b31b-447158a62875; detaching it from the instance and deleting it from the info cache [ 1097.267387] env[62600]: DEBUG nova.network.neutron [req-7723df6b-39d2-45b3-a58d-a70cccf0cf8f req-77d6360a-5c6e-4233-b978-cf2a45f99066 service nova] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.422299] env[62600]: DEBUG oslo_vmware.api [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.727662] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.728014] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.735699] env[62600]: DEBUG nova.network.neutron [-] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.770029] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e838b8e0-5bf6-4ba6-825c-91b95fa6f6b6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.779855] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144ad182-31ef-4aa6-8921-385e25b73f82 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.807585] env[62600]: DEBUG nova.compute.manager [req-7723df6b-39d2-45b3-a58d-a70cccf0cf8f req-77d6360a-5c6e-4233-b978-cf2a45f99066 service nova] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Detach interface failed, port_id=028950a6-9715-4e67-b31b-447158a62875, reason: Instance 8417575c-d68d-464f-8c22-ab358deb76d8 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1097.922800] env[62600]: DEBUG oslo_vmware.api [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.231199] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.232034] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.232034] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.232178] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62600) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1098.232965] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41414731-839e-48aa-b795-2aa0ffd3d909 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.240400] env[62600]: INFO nova.compute.manager [-] [instance: 8417575c-d68d-464f-8c22-ab358deb76d8] Took 1.45 seconds to deallocate network for instance. [ 1098.243262] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f80fff61-4cff-4a7b-b73d-dd8ef0c147ab {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.261976] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1026871a-fb68-4e3f-843f-f8c6343f3cce {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.269179] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-358a136c-53d3-419f-84d3-0b2b71192cfc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.299989] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180010MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=62600) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1098.300407] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.300735] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.422371] env[62600]: DEBUG oslo_vmware.api [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.752414] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.927951] env[62600]: DEBUG oslo_vmware.api [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.329646] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance c2f38d56-72fa-4bf0-a2cb-13d666571466 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1099.329916] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance eb3c5624-0599-4e81-b056-bedf34329646 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1099.330158] env[62600]: WARNING nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 8417575c-d68d-464f-8c22-ab358deb76d8 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1099.330340] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 692fa608-39f4-4d7d-9f30-b6559878b970 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1099.330531] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 1817183b-44a5-41ec-a2bf-818edee264b3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1099.330786] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1099.330953] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1099.411293] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-269465a7-968e-4cfa-b730-e3e5b60f356f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.420389] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de119e7e-82bd-4827-8b16-e14af5b4bf5a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.426504] env[62600]: DEBUG oslo_vmware.api [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.453460] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-625698fd-ab0c-4382-8c9b-45117e9b5218 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.461286] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a837f253-b952-4d07-8f30-6c03c6b4ba54 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.475170] env[62600]: DEBUG nova.compute.provider_tree [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1099.924279] env[62600]: DEBUG oslo_vmware.api [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.978498] env[62600]: DEBUG nova.scheduler.client.report [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1100.427438] env[62600]: DEBUG oslo_vmware.api [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.483631] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62600) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1100.483917] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.183s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.484289] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.732s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.484571] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.486956] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1100.487167] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Cleaning up deleted instances {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1100.507016] env[62600]: INFO nova.scheduler.client.report [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Deleted allocations for instance 8417575c-d68d-464f-8c22-ab358deb76d8 [ 1100.928679] env[62600]: DEBUG oslo_vmware.api [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.005510] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] There are 55 instances to clean {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1101.005760] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 1d6bc00e-b18b-4333-b500-bb9d9d00d4e8] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1101.018358] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ea900bf-0acd-4578-98e0-72538fc91314 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "8417575c-d68d-464f-8c22-ab358deb76d8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.358s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.426422] env[62600]: DEBUG oslo_vmware.api [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223049, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.509401] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 9c7a847e-24e4-4ea7-b9b9-0c257ad5f7b7] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1101.927914] env[62600]: DEBUG oslo_vmware.api [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223049, 'name': ReconfigVM_Task, 'duration_secs': 5.760075} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.928172] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.928425] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Reconfigured VM to detach interface {{(pid=62600) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1102.014957] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 4607dacd-28ab-4173-ba86-8a78bec19dbc] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1102.518945] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 0416ef03-fd45-405f-9047-76b34a839217] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1102.645247] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "80a9de40-675d-456b-a91b-432d2315144b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.645510] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "80a9de40-675d-456b-a91b-432d2315144b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.022669] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: a20222a7-f29b-468f-8105-50ff9e31bab2] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1103.148189] env[62600]: DEBUG nova.compute.manager [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1103.284663] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.284964] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquired lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.285232] env[62600]: DEBUG nova.network.neutron [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1103.526669] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 5d7c2a17-9dee-407e-a7ee-6bb88f56905f] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1103.670801] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.671073] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.672726] env[62600]: INFO nova.compute.claims [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1103.699628] env[62600]: DEBUG oslo_vmware.rw_handles [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5232e42b-2732-3783-2af2-e77e3774d9af/disk-0.vmdk. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1103.700788] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70030b2c-6a43-4217-b9ca-55162a759cb4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.708508] env[62600]: DEBUG oslo_vmware.rw_handles [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5232e42b-2732-3783-2af2-e77e3774d9af/disk-0.vmdk is in state: ready. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1103.708685] env[62600]: ERROR oslo_vmware.rw_handles [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5232e42b-2732-3783-2af2-e77e3774d9af/disk-0.vmdk due to incomplete transfer. [ 1103.708904] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9e5f46a2-27a6-458b-8a20-e94f6b3af375 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.716086] env[62600]: DEBUG oslo_vmware.rw_handles [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5232e42b-2732-3783-2af2-e77e3774d9af/disk-0.vmdk. {{(pid=62600) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1103.716284] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Uploaded image 10f7ff7d-b198-418a-8817-0b212a53d0a9 to the Glance image server {{(pid=62600) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1103.718561] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Destroying the VM {{(pid=62600) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1103.718793] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-29dc45ee-297c-4d4e-b738-91f857c04d1b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.724770] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1103.724770] env[62600]: value = "task-1223053" [ 1103.724770] env[62600]: _type = "Task" [ 1103.724770] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.732128] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223053, 'name': Destroy_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.782826] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "1817183b-44a5-41ec-a2bf-818edee264b3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.783106] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "1817183b-44a5-41ec-a2bf-818edee264b3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.783517] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "1817183b-44a5-41ec-a2bf-818edee264b3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.783517] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "1817183b-44a5-41ec-a2bf-818edee264b3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.783715] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "1817183b-44a5-41ec-a2bf-818edee264b3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.786101] env[62600]: INFO nova.compute.manager [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Terminating instance [ 1103.789627] env[62600]: DEBUG nova.compute.manager [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1103.789836] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1103.790763] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-410f27ef-d3b4-45a1-bad3-adbefb0911cd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.799057] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1103.799057] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf1b3455-188e-4d6e-9233-e3c77a45b15f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.804014] env[62600]: DEBUG oslo_vmware.api [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1103.804014] env[62600]: value = "task-1223054" [ 1103.804014] env[62600]: _type = "Task" [ 1103.804014] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.812209] env[62600]: DEBUG oslo_vmware.api [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223054, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.005950] env[62600]: INFO nova.network.neutron [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Port 1e67e320-f1e0-40f7-b0df-9cac23dbf102 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1104.006470] env[62600]: DEBUG nova.network.neutron [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Updating instance_info_cache with network_info: [{"id": "5796bb3c-0926-47b4-bcdb-c54c18f83200", "address": "fa:16:3e:94:99:6b", "network": {"id": "8ff5e8fa-edd8-40c9-94cb-f780ce4d08bb", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-617100721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b028450f2da445fb83e37adfc86bba68", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "184687d6-125a-4b58-bb5b-fdb404088eda", "external-id": "nsx-vlan-transportzone-134", "segmentation_id": 134, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5796bb3c-09", "ovs_interfaceid": "5796bb3c-0926-47b4-bcdb-c54c18f83200", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.030189] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: aff91190-e19b-4f1c-8b47-9539360a4596] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1104.234532] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223053, 'name': Destroy_Task, 'duration_secs': 0.300602} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.234808] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Destroyed the VM [ 1104.235098] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Deleting Snapshot of the VM instance {{(pid=62600) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1104.235353] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-86cdf404-720d-44e4-829c-8b036e22aad5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.240804] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1104.240804] env[62600]: value = "task-1223055" [ 1104.240804] env[62600]: _type = "Task" [ 1104.240804] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.248495] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223055, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.314746] env[62600]: DEBUG oslo_vmware.api [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223054, 'name': PowerOffVM_Task, 'duration_secs': 0.259849} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.315083] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1104.315318] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1104.315605] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-73c8ff25-5bc3-4f07-ad52-91d22f13c0a9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.371203] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1104.371467] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1104.371677] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Deleting the datastore file [datastore1] 1817183b-44a5-41ec-a2bf-818edee264b3 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1104.371996] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b66efc17-04bc-4d9d-8809-0084491a83ff {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.379927] env[62600]: DEBUG oslo_vmware.api [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1104.379927] env[62600]: value = "task-1223057" [ 1104.379927] env[62600]: _type = "Task" [ 1104.379927] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.387943] env[62600]: DEBUG oslo_vmware.api [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223057, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.509421] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Releasing lock "refresh_cache-1817183b-44a5-41ec-a2bf-818edee264b3" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.533192] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: e4bc0326-66d8-45e2-bcd8-11c4fab16e34] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1104.746875] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba53f3b-62e2-4855-bbc1-d5331cfe17c3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.756361] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-852f7941-b962-4ca3-9dc4-9deeb97daed8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.759531] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223055, 'name': RemoveSnapshot_Task, 'duration_secs': 0.37011} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.759791] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Deleted Snapshot of the VM instance {{(pid=62600) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1104.760082] env[62600]: DEBUG nova.compute.manager [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1104.761090] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35a8dab-acfd-485b-9e90-385397628c02 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.786935] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c79fa7-44ba-4cce-8f1f-6c7a5469e2c1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.796456] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e4aa1b5-a4f2-4e0b-a7c1-f44a6e2522bc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.808892] env[62600]: DEBUG nova.compute.provider_tree [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1104.890623] env[62600]: DEBUG oslo_vmware.api [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223057, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196264} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.890911] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1104.891089] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1104.891276] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1104.891454] env[62600]: INFO nova.compute.manager [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1104.891697] env[62600]: DEBUG oslo.service.loopingcall [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1104.891882] env[62600]: DEBUG nova.compute.manager [-] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1104.891978] env[62600]: DEBUG nova.network.neutron [-] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1105.013943] env[62600]: DEBUG oslo_concurrency.lockutils [None req-3e888e5e-a85e-46c5-85c4-e16ad6f0388f tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "interface-1817183b-44a5-41ec-a2bf-818edee264b3-1e67e320-f1e0-40f7-b0df-9cac23dbf102" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.668s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.036251] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 310a9093-4066-4bc3-af94-5d606860dbad] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1105.296093] env[62600]: INFO nova.compute.manager [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Shelve offloading [ 1105.298296] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1105.298546] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d98755c4-4a5c-4877-b36c-27e379949a53 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.304812] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1105.304812] env[62600]: value = "task-1223058" [ 1105.304812] env[62600]: _type = "Task" [ 1105.304812] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.312841] env[62600]: DEBUG nova.scheduler.client.report [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1105.315674] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223058, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.503912] env[62600]: DEBUG nova.compute.manager [req-a71b01fa-4266-414f-a57b-9c8400fdaaad req-2e8ff62d-1b68-46cf-8176-6192fde406cd service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Received event network-vif-deleted-5796bb3c-0926-47b4-bcdb-c54c18f83200 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1105.504142] env[62600]: INFO nova.compute.manager [req-a71b01fa-4266-414f-a57b-9c8400fdaaad req-2e8ff62d-1b68-46cf-8176-6192fde406cd service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Neutron deleted interface 5796bb3c-0926-47b4-bcdb-c54c18f83200; detaching it from the instance and deleting it from the info cache [ 1105.504319] env[62600]: DEBUG nova.network.neutron [req-a71b01fa-4266-414f-a57b-9c8400fdaaad req-2e8ff62d-1b68-46cf-8176-6192fde406cd service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.539539] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 02c6dac9-535b-436b-bbee-c8031cb5d7f5] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1105.815295] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] VM already powered off {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1105.815540] env[62600]: DEBUG nova.compute.manager [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1105.816358] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f38a699-57b3-4509-92fd-a152b2fd5f0b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.819171] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.148s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.819652] env[62600]: DEBUG nova.compute.manager [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1105.827749] env[62600]: DEBUG oslo_concurrency.lockutils [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.827749] env[62600]: DEBUG oslo_concurrency.lockutils [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquired lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.827749] env[62600]: DEBUG nova.network.neutron [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1105.984546] env[62600]: DEBUG nova.network.neutron [-] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.006952] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aee4e6ef-a10a-4b9c-9fe6-0a316784aa4e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.016184] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fa98d92-bc12-41d9-b2dc-d4a5b2175e8e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.040028] env[62600]: DEBUG nova.compute.manager [req-a71b01fa-4266-414f-a57b-9c8400fdaaad req-2e8ff62d-1b68-46cf-8176-6192fde406cd service nova] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Detach interface failed, port_id=5796bb3c-0926-47b4-bcdb-c54c18f83200, reason: Instance 1817183b-44a5-41ec-a2bf-818edee264b3 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1106.043632] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 4d12c133-84d7-4ce9-83db-e4b70b8947f3] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1106.324824] env[62600]: DEBUG nova.compute.utils [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1106.329766] env[62600]: DEBUG nova.compute.manager [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1106.329766] env[62600]: DEBUG nova.network.neutron [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1106.366661] env[62600]: DEBUG nova.policy [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '94751040d0fd41a895010c0f2f019131', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef3bb131f3994f7a8673dafff6d353c1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 1106.487438] env[62600]: INFO nova.compute.manager [-] [instance: 1817183b-44a5-41ec-a2bf-818edee264b3] Took 1.60 seconds to deallocate network for instance. [ 1106.526489] env[62600]: DEBUG nova.network.neutron [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Updating instance_info_cache with network_info: [{"id": "280baabd-3083-4f33-b294-1f988193665b", "address": "fa:16:3e:72:db:a2", "network": {"id": "dc21424f-b891-428d-a74e-76ab78d757a8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1428037886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "952323d62f034126a9b46750d03fe6ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap280baabd-30", "ovs_interfaceid": "280baabd-3083-4f33-b294-1f988193665b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.547047] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 39deb498-6bf0-4f3b-932b-8068fc48271e] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1106.647749] env[62600]: DEBUG nova.network.neutron [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Successfully created port: 4cf97208-acde-4474-94d5-27c1ed6ddd8c {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1106.828134] env[62600]: DEBUG nova.compute.manager [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1106.995813] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.996165] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.996403] env[62600]: DEBUG nova.objects.instance [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lazy-loading 'resources' on Instance uuid 1817183b-44a5-41ec-a2bf-818edee264b3 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1107.029413] env[62600]: DEBUG oslo_concurrency.lockutils [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Releasing lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1107.050211] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 234dc7f8-e2ef-4f20-b25b-dcd61ce20c57] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1107.242525] env[62600]: DEBUG nova.compute.manager [req-79925802-9076-4806-85c4-6a225692e663 req-c68f3c20-9c63-403b-9712-76f86ad5e6d9 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Received event network-vif-unplugged-280baabd-3083-4f33-b294-1f988193665b {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1107.243015] env[62600]: DEBUG oslo_concurrency.lockutils [req-79925802-9076-4806-85c4-6a225692e663 req-c68f3c20-9c63-403b-9712-76f86ad5e6d9 service nova] Acquiring lock "692fa608-39f4-4d7d-9f30-b6559878b970-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.243193] env[62600]: DEBUG oslo_concurrency.lockutils [req-79925802-9076-4806-85c4-6a225692e663 req-c68f3c20-9c63-403b-9712-76f86ad5e6d9 service nova] Lock "692fa608-39f4-4d7d-9f30-b6559878b970-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.243382] env[62600]: DEBUG oslo_concurrency.lockutils [req-79925802-9076-4806-85c4-6a225692e663 req-c68f3c20-9c63-403b-9712-76f86ad5e6d9 service nova] Lock "692fa608-39f4-4d7d-9f30-b6559878b970-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.243508] env[62600]: DEBUG nova.compute.manager [req-79925802-9076-4806-85c4-6a225692e663 req-c68f3c20-9c63-403b-9712-76f86ad5e6d9 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] No waiting events found dispatching network-vif-unplugged-280baabd-3083-4f33-b294-1f988193665b {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1107.243687] env[62600]: WARNING nova.compute.manager [req-79925802-9076-4806-85c4-6a225692e663 req-c68f3c20-9c63-403b-9712-76f86ad5e6d9 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Received unexpected event network-vif-unplugged-280baabd-3083-4f33-b294-1f988193665b for instance with vm_state shelved and task_state shelving_offloading. [ 1107.339464] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1107.340839] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2cc73f-811f-48b2-96f8-55312b7413c6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.350350] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1107.350613] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3020054f-758f-453f-9ded-8d8f4626d586 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.412418] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1107.412701] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1107.412946] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Deleting the datastore file [datastore1] 692fa608-39f4-4d7d-9f30-b6559878b970 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1107.413224] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f04d0157-b1f3-405d-a622-b17209da7af7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.419432] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1107.419432] env[62600]: value = "task-1223060" [ 1107.419432] env[62600]: _type = "Task" [ 1107.419432] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.427193] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223060, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.553294] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 81333628-2616-4557-88d1-0e7164e1b16d] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1107.674381] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06d2650c-3c63-4bc1-9452-2f03d9f8362a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.681661] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff29c651-7f01-48be-9608-426dc8c90d14 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.712044] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e2e727e-9326-4619-be46-8f16c6031835 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.719037] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56118605-fd4f-4a3b-8e46-b383dcbd3035 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.733416] env[62600]: DEBUG nova.compute.provider_tree [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1107.846996] env[62600]: DEBUG nova.compute.manager [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1107.871741] env[62600]: DEBUG nova.virt.hardware [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1107.872015] env[62600]: DEBUG nova.virt.hardware [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1107.872192] env[62600]: DEBUG nova.virt.hardware [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1107.872381] env[62600]: DEBUG nova.virt.hardware [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1107.872535] env[62600]: DEBUG nova.virt.hardware [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1107.872684] env[62600]: DEBUG nova.virt.hardware [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1107.872949] env[62600]: DEBUG nova.virt.hardware [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1107.873145] env[62600]: DEBUG nova.virt.hardware [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1107.873325] env[62600]: DEBUG nova.virt.hardware [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1107.873498] env[62600]: DEBUG nova.virt.hardware [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1107.873678] env[62600]: DEBUG nova.virt.hardware [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1107.874866] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-489a82ce-b46e-4a41-b33d-14548536abe7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.882860] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a46f5973-c701-44c9-ae9e-34d42a50fa06 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.927794] env[62600]: DEBUG oslo_vmware.api [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223060, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.120644} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.928062] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1107.928262] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1107.928445] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1107.946157] env[62600]: INFO nova.scheduler.client.report [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Deleted allocations for instance 692fa608-39f4-4d7d-9f30-b6559878b970 [ 1108.015688] env[62600]: DEBUG nova.compute.manager [req-dfc457b1-be01-41fd-88ba-f8c7e2d8a17e req-8f63f6ee-0e65-4eeb-9ade-b93a62a8b2c7 service nova] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Received event network-vif-plugged-4cf97208-acde-4474-94d5-27c1ed6ddd8c {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1108.015955] env[62600]: DEBUG oslo_concurrency.lockutils [req-dfc457b1-be01-41fd-88ba-f8c7e2d8a17e req-8f63f6ee-0e65-4eeb-9ade-b93a62a8b2c7 service nova] Acquiring lock "80a9de40-675d-456b-a91b-432d2315144b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.016220] env[62600]: DEBUG oslo_concurrency.lockutils [req-dfc457b1-be01-41fd-88ba-f8c7e2d8a17e req-8f63f6ee-0e65-4eeb-9ade-b93a62a8b2c7 service nova] Lock "80a9de40-675d-456b-a91b-432d2315144b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.016402] env[62600]: DEBUG oslo_concurrency.lockutils [req-dfc457b1-be01-41fd-88ba-f8c7e2d8a17e req-8f63f6ee-0e65-4eeb-9ade-b93a62a8b2c7 service nova] Lock "80a9de40-675d-456b-a91b-432d2315144b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.016580] env[62600]: DEBUG nova.compute.manager [req-dfc457b1-be01-41fd-88ba-f8c7e2d8a17e req-8f63f6ee-0e65-4eeb-9ade-b93a62a8b2c7 service nova] [instance: 80a9de40-675d-456b-a91b-432d2315144b] No waiting events found dispatching network-vif-plugged-4cf97208-acde-4474-94d5-27c1ed6ddd8c {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1108.016750] env[62600]: WARNING nova.compute.manager [req-dfc457b1-be01-41fd-88ba-f8c7e2d8a17e req-8f63f6ee-0e65-4eeb-9ade-b93a62a8b2c7 service nova] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Received unexpected event network-vif-plugged-4cf97208-acde-4474-94d5-27c1ed6ddd8c for instance with vm_state building and task_state spawning. [ 1108.056348] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 784e6d19-36ef-4c01-9e9f-5a083ed08608] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1108.102787] env[62600]: DEBUG nova.network.neutron [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Successfully updated port: 4cf97208-acde-4474-94d5-27c1ed6ddd8c {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1108.236248] env[62600]: DEBUG nova.scheduler.client.report [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1108.451067] env[62600]: DEBUG oslo_concurrency.lockutils [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.560037] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 84dc284d-c7f4-4fc2-b539-8d4820429b02] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1108.607820] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "refresh_cache-80a9de40-675d-456b-a91b-432d2315144b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.607974] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquired lock "refresh_cache-80a9de40-675d-456b-a91b-432d2315144b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.608140] env[62600]: DEBUG nova.network.neutron [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1108.741495] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.745s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.744019] env[62600]: DEBUG oslo_concurrency.lockutils [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.293s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.744019] env[62600]: DEBUG nova.objects.instance [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lazy-loading 'resources' on Instance uuid 692fa608-39f4-4d7d-9f30-b6559878b970 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1108.758548] env[62600]: INFO nova.scheduler.client.report [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Deleted allocations for instance 1817183b-44a5-41ec-a2bf-818edee264b3 [ 1109.062874] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: f18df958-2885-4d8c-a9ad-140faaed1178] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1109.141127] env[62600]: DEBUG nova.network.neutron [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1109.246661] env[62600]: DEBUG nova.objects.instance [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lazy-loading 'numa_topology' on Instance uuid 692fa608-39f4-4d7d-9f30-b6559878b970 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1109.266703] env[62600]: DEBUG oslo_concurrency.lockutils [None req-d3ecabc8-2889-4a32-abfe-8ff9ab2a9e2e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "1817183b-44a5-41ec-a2bf-818edee264b3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.484s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.270526] env[62600]: DEBUG nova.compute.manager [req-0d0b64f6-8407-4b36-bf46-7124b719f38d req-e5daa6f4-567e-464f-ae24-23571ce06577 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Received event network-changed-280baabd-3083-4f33-b294-1f988193665b {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1109.270625] env[62600]: DEBUG nova.compute.manager [req-0d0b64f6-8407-4b36-bf46-7124b719f38d req-e5daa6f4-567e-464f-ae24-23571ce06577 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Refreshing instance network info cache due to event network-changed-280baabd-3083-4f33-b294-1f988193665b. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1109.271318] env[62600]: DEBUG oslo_concurrency.lockutils [req-0d0b64f6-8407-4b36-bf46-7124b719f38d req-e5daa6f4-567e-464f-ae24-23571ce06577 service nova] Acquiring lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.271318] env[62600]: DEBUG oslo_concurrency.lockutils [req-0d0b64f6-8407-4b36-bf46-7124b719f38d req-e5daa6f4-567e-464f-ae24-23571ce06577 service nova] Acquired lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.271318] env[62600]: DEBUG nova.network.neutron [req-0d0b64f6-8407-4b36-bf46-7124b719f38d req-e5daa6f4-567e-464f-ae24-23571ce06577 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Refreshing network info cache for port 280baabd-3083-4f33-b294-1f988193665b {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1109.276159] env[62600]: DEBUG nova.network.neutron [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Updating instance_info_cache with network_info: [{"id": "4cf97208-acde-4474-94d5-27c1ed6ddd8c", "address": "fa:16:3e:eb:1c:12", "network": {"id": "e7547ac4-24a6-4778-a17a-40e79e7a0d0c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-320671681-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3bb131f3994f7a8673dafff6d353c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4cf97208-ac", "ovs_interfaceid": "4cf97208-acde-4474-94d5-27c1ed6ddd8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.566056] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 88cf521d-f4a3-493a-894d-c7f799dc0443] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1109.748863] env[62600]: DEBUG nova.objects.base [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Object Instance<692fa608-39f4-4d7d-9f30-b6559878b970> lazy-loaded attributes: resources,numa_topology {{(pid=62600) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1109.778512] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Releasing lock "refresh_cache-80a9de40-675d-456b-a91b-432d2315144b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.778644] env[62600]: DEBUG nova.compute.manager [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Instance network_info: |[{"id": "4cf97208-acde-4474-94d5-27c1ed6ddd8c", "address": "fa:16:3e:eb:1c:12", "network": {"id": "e7547ac4-24a6-4778-a17a-40e79e7a0d0c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-320671681-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3bb131f3994f7a8673dafff6d353c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4cf97208-ac", "ovs_interfaceid": "4cf97208-acde-4474-94d5-27c1ed6ddd8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1109.781481] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:eb:1c:12', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4cf97208-acde-4474-94d5-27c1ed6ddd8c', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1109.789186] env[62600]: DEBUG oslo.service.loopingcall [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1109.789604] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1109.789784] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c5ff4161-64ec-4bd7-a598-9e0e284c48f8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.811843] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1109.811843] env[62600]: value = "task-1223061" [ 1109.811843] env[62600]: _type = "Task" [ 1109.811843] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.820691] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1223061, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.830712] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-601b82df-3d22-41bb-ade8-4ccf1816abe3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.837386] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82f4b899-5205-4b7b-a407-e65255b26a4e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.871171] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95770be8-4da2-4c73-a59a-b7489d4f0aaa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.883086] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb634a5e-a630-4459-9f25-04e184396b18 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.897966] env[62600]: DEBUG nova.compute.provider_tree [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1110.034866] env[62600]: DEBUG nova.network.neutron [req-0d0b64f6-8407-4b36-bf46-7124b719f38d req-e5daa6f4-567e-464f-ae24-23571ce06577 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Updated VIF entry in instance network info cache for port 280baabd-3083-4f33-b294-1f988193665b. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1110.035296] env[62600]: DEBUG nova.network.neutron [req-0d0b64f6-8407-4b36-bf46-7124b719f38d req-e5daa6f4-567e-464f-ae24-23571ce06577 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Updating instance_info_cache with network_info: [{"id": "280baabd-3083-4f33-b294-1f988193665b", "address": "fa:16:3e:72:db:a2", "network": {"id": "dc21424f-b891-428d-a74e-76ab78d757a8", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1428037886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "952323d62f034126a9b46750d03fe6ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap280baabd-30", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.042786] env[62600]: DEBUG nova.compute.manager [req-08decbd2-da18-4e5e-af01-bf9bb9f04f8a req-b27fab42-456c-480a-98d4-6e8a08cc431d service nova] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Received event network-changed-4cf97208-acde-4474-94d5-27c1ed6ddd8c {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1110.043024] env[62600]: DEBUG nova.compute.manager [req-08decbd2-da18-4e5e-af01-bf9bb9f04f8a req-b27fab42-456c-480a-98d4-6e8a08cc431d service nova] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Refreshing instance network info cache due to event network-changed-4cf97208-acde-4474-94d5-27c1ed6ddd8c. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1110.043303] env[62600]: DEBUG oslo_concurrency.lockutils [req-08decbd2-da18-4e5e-af01-bf9bb9f04f8a req-b27fab42-456c-480a-98d4-6e8a08cc431d service nova] Acquiring lock "refresh_cache-80a9de40-675d-456b-a91b-432d2315144b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.043492] env[62600]: DEBUG oslo_concurrency.lockutils [req-08decbd2-da18-4e5e-af01-bf9bb9f04f8a req-b27fab42-456c-480a-98d4-6e8a08cc431d service nova] Acquired lock "refresh_cache-80a9de40-675d-456b-a91b-432d2315144b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.043683] env[62600]: DEBUG nova.network.neutron [req-08decbd2-da18-4e5e-af01-bf9bb9f04f8a req-b27fab42-456c-480a-98d4-6e8a08cc431d service nova] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Refreshing network info cache for port 4cf97208-acde-4474-94d5-27c1ed6ddd8c {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1110.068649] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 248a5371-6ff7-4da7-ae97-d638bea0123a] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1110.322017] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1223061, 'name': CreateVM_Task, 'duration_secs': 0.300391} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.322200] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1110.322880] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.323067] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.323410] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1110.323656] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54e63494-3e29-4332-9589-0dc5092c10a2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.328154] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1110.328154] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528844ba-70df-34ed-1222-9acbdc442f2d" [ 1110.328154] env[62600]: _type = "Task" [ 1110.328154] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.336056] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528844ba-70df-34ed-1222-9acbdc442f2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.379169] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "eb3c5624-0599-4e81-b056-bedf34329646" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.379432] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "eb3c5624-0599-4e81-b056-bedf34329646" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.379646] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "eb3c5624-0599-4e81-b056-bedf34329646-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.379835] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "eb3c5624-0599-4e81-b056-bedf34329646-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.380068] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "eb3c5624-0599-4e81-b056-bedf34329646-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.382119] env[62600]: INFO nova.compute.manager [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Terminating instance [ 1110.383938] env[62600]: DEBUG nova.compute.manager [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1110.384161] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1110.385011] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd67707-5ebc-40d5-a6ee-f052c86a1324 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.393052] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1110.393052] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b4ce6928-b29a-459a-ae75-0338b60bf4d0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.399877] env[62600]: DEBUG nova.scheduler.client.report [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1110.402993] env[62600]: DEBUG oslo_vmware.api [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1110.402993] env[62600]: value = "task-1223062" [ 1110.402993] env[62600]: _type = "Task" [ 1110.402993] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.412397] env[62600]: DEBUG oslo_vmware.api [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223062, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.422013] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "692fa608-39f4-4d7d-9f30-b6559878b970" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.538441] env[62600]: DEBUG oslo_concurrency.lockutils [req-0d0b64f6-8407-4b36-bf46-7124b719f38d req-e5daa6f4-567e-464f-ae24-23571ce06577 service nova] Releasing lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.571817] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 6447a26d-fc30-41eb-bd42-94341a86c2e8] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1110.729917] env[62600]: DEBUG nova.network.neutron [req-08decbd2-da18-4e5e-af01-bf9bb9f04f8a req-b27fab42-456c-480a-98d4-6e8a08cc431d service nova] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Updated VIF entry in instance network info cache for port 4cf97208-acde-4474-94d5-27c1ed6ddd8c. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1110.730412] env[62600]: DEBUG nova.network.neutron [req-08decbd2-da18-4e5e-af01-bf9bb9f04f8a req-b27fab42-456c-480a-98d4-6e8a08cc431d service nova] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Updating instance_info_cache with network_info: [{"id": "4cf97208-acde-4474-94d5-27c1ed6ddd8c", "address": "fa:16:3e:eb:1c:12", "network": {"id": "e7547ac4-24a6-4778-a17a-40e79e7a0d0c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-320671681-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3bb131f3994f7a8673dafff6d353c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4cf97208-ac", "ovs_interfaceid": "4cf97208-acde-4474-94d5-27c1ed6ddd8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.837763] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528844ba-70df-34ed-1222-9acbdc442f2d, 'name': SearchDatastore_Task, 'duration_secs': 0.011849} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.838182] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.838426] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1110.838669] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.838825] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.839015] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1110.839273] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-785e7fc1-4412-4be2-bd47-7d14c5c0ccb5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.847008] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1110.847197] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1110.847852] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb3b45f3-43d0-45a5-8b94-aa1abb6f5553 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.853642] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1110.853642] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f9d10e-57ce-2884-4cb4-fee6ebd712dd" [ 1110.853642] env[62600]: _type = "Task" [ 1110.853642] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.859156] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f9d10e-57ce-2884-4cb4-fee6ebd712dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.904497] env[62600]: DEBUG oslo_concurrency.lockutils [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.161s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.914007] env[62600]: DEBUG oslo_vmware.api [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223062, 'name': PowerOffVM_Task, 'duration_secs': 0.181462} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.914269] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1110.914451] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1110.914664] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a136dc19-a6cd-4794-905b-67abedae8d67 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.974365] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1110.974642] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1110.974857] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Deleting the datastore file [datastore1] eb3c5624-0599-4e81-b056-bedf34329646 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1110.975143] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de5f329c-e2ea-4df5-a71c-9c9acde3bb0d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.982115] env[62600]: DEBUG oslo_vmware.api [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for the task: (returnval){ [ 1110.982115] env[62600]: value = "task-1223064" [ 1110.982115] env[62600]: _type = "Task" [ 1110.982115] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.989704] env[62600]: DEBUG oslo_vmware.api [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223064, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.075646] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 9ed08410-a4cb-43c4-9271-2c04d9c87eac] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1111.232685] env[62600]: DEBUG oslo_concurrency.lockutils [req-08decbd2-da18-4e5e-af01-bf9bb9f04f8a req-b27fab42-456c-480a-98d4-6e8a08cc431d service nova] Releasing lock "refresh_cache-80a9de40-675d-456b-a91b-432d2315144b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.363585] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f9d10e-57ce-2884-4cb4-fee6ebd712dd, 'name': SearchDatastore_Task, 'duration_secs': 0.008076} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.364332] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7cdcda3a-c09a-4a0b-bc2e-2d79c6f57071 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.369408] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1111.369408] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c3220f-b5e3-b183-d633-1f0e558caee5" [ 1111.369408] env[62600]: _type = "Task" [ 1111.369408] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.376431] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c3220f-b5e3-b183-d633-1f0e558caee5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.414159] env[62600]: DEBUG oslo_concurrency.lockutils [None req-62ce73e6-d087-4d7d-b2f0-0ae1faf7736c tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "692fa608-39f4-4d7d-9f30-b6559878b970" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 18.984s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.415042] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "692fa608-39f4-4d7d-9f30-b6559878b970" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.993s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.415232] env[62600]: INFO nova.compute.manager [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Unshelving [ 1111.492329] env[62600]: DEBUG oslo_vmware.api [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Task: {'id': task-1223064, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144842} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.492594] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1111.492796] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1111.493028] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1111.493221] env[62600]: INFO nova.compute.manager [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1111.493466] env[62600]: DEBUG oslo.service.loopingcall [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1111.493656] env[62600]: DEBUG nova.compute.manager [-] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1111.493755] env[62600]: DEBUG nova.network.neutron [-] [instance: eb3c5624-0599-4e81-b056-bedf34329646] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1111.578777] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 043ac602-6c4a-4275-b019-5d717579e7c0] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1111.880167] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52c3220f-b5e3-b183-d633-1f0e558caee5, 'name': SearchDatastore_Task, 'duration_secs': 0.009225} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.880451] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.880716] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 80a9de40-675d-456b-a91b-432d2315144b/80a9de40-675d-456b-a91b-432d2315144b.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1111.880998] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4045f067-3c0e-4ef9-bfa1-573f279f89ba {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.888842] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1111.888842] env[62600]: value = "task-1223065" [ 1111.888842] env[62600]: _type = "Task" [ 1111.888842] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.897208] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223065, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.072809] env[62600]: DEBUG nova.compute.manager [req-e988d444-69d0-4c3e-9b7d-ea933945445f req-0cbefc81-9f67-4d2e-a576-3219c9cadfdd service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Received event network-vif-deleted-b8e9cd2e-1990-4997-b1fa-cb4411d39fde {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1112.072994] env[62600]: INFO nova.compute.manager [req-e988d444-69d0-4c3e-9b7d-ea933945445f req-0cbefc81-9f67-4d2e-a576-3219c9cadfdd service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Neutron deleted interface b8e9cd2e-1990-4997-b1fa-cb4411d39fde; detaching it from the instance and deleting it from the info cache [ 1112.073245] env[62600]: DEBUG nova.network.neutron [req-e988d444-69d0-4c3e-9b7d-ea933945445f req-0cbefc81-9f67-4d2e-a576-3219c9cadfdd service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.082254] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 4eaa22b9-5834-4425-82a9-d16e316f3a52] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1112.224514] env[62600]: DEBUG nova.network.neutron [-] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.399345] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223065, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.46682} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.399628] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 80a9de40-675d-456b-a91b-432d2315144b/80a9de40-675d-456b-a91b-432d2315144b.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1112.399831] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1112.400109] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d7cbd9ad-07ae-47a7-9a98-9fa5c7d665b7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.406586] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1112.406586] env[62600]: value = "task-1223066" [ 1112.406586] env[62600]: _type = "Task" [ 1112.406586] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.415077] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223066, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.426876] env[62600]: DEBUG nova.compute.utils [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1112.576568] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0a5f7936-2453-4422-a238-d5fc157ddfc6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.585469] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 48bbbdaf-496d-4780-b467-6b393257535d] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1112.593039] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67ecf152-bc89-4fbe-9184-d5fb8441ca1c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.628552] env[62600]: DEBUG nova.compute.manager [req-e988d444-69d0-4c3e-9b7d-ea933945445f req-0cbefc81-9f67-4d2e-a576-3219c9cadfdd service nova] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Detach interface failed, port_id=b8e9cd2e-1990-4997-b1fa-cb4411d39fde, reason: Instance eb3c5624-0599-4e81-b056-bedf34329646 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1112.727244] env[62600]: INFO nova.compute.manager [-] [instance: eb3c5624-0599-4e81-b056-bedf34329646] Took 1.23 seconds to deallocate network for instance. [ 1112.919025] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223066, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087117} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.919025] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1112.919025] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd11474-6971-47fd-b5c8-10b8d0c658f8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.937160] env[62600]: INFO nova.virt.block_device [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Booting with volume 34095f5d-35c4-4bad-88a3-e3e627c33b5c at /dev/sdb [ 1112.958119] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 80a9de40-675d-456b-a91b-432d2315144b/80a9de40-675d-456b-a91b-432d2315144b.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1112.958765] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60e20c84-6d44-493d-9e12-9c922a7df738 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.987975] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1112.987975] env[62600]: value = "task-1223067" [ 1112.987975] env[62600]: _type = "Task" [ 1112.987975] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.998989] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223067, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.025045] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3ec6db93-9be8-4d09-9f47-14e1d1c7d51f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.033577] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece423d7-c256-4a04-923d-bf1eeba6f3ae {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.058899] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cd772bc9-6cd8-4263-968d-e089abd98cef {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.066502] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1acd113-9d8a-4b4f-b18b-845352da2d58 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.092931] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d001631-97d9-480b-a1b5-b0bc35f834eb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.099468] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb106a1-e680-4ef2-b6da-7ce4a70726d1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.112883] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: d2e43ba2-4ccf-4625-91ba-78a6f2632461] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1113.114960] env[62600]: DEBUG nova.virt.block_device [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Updating existing volume attachment record: 8d50d655-28eb-47a6-93fa-14394369ca57 {{(pid=62600) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1113.233795] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.234106] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.234356] env[62600]: DEBUG nova.objects.instance [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lazy-loading 'resources' on Instance uuid eb3c5624-0599-4e81-b056-bedf34329646 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1113.498184] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223067, 'name': ReconfigVM_Task, 'duration_secs': 0.422956} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.498449] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 80a9de40-675d-456b-a91b-432d2315144b/80a9de40-675d-456b-a91b-432d2315144b.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1113.499050] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c59652ec-bac7-4cea-a981-f17e2969e90e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.505666] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1113.505666] env[62600]: value = "task-1223071" [ 1113.505666] env[62600]: _type = "Task" [ 1113.505666] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.513202] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223071, 'name': Rename_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.620398] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: d627a701-77cd-4a1d-9e52-0fc4adc93391] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1113.800008] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc28e59-069b-44f7-aad0-6a33fb4f8e26 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.807233] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd94ef3c-4759-40d4-8d29-a78b586a5d5e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.838020] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff146b8-1c5d-480b-91d2-3d8527116def {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.845046] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1efbcd8c-f3f7-4329-950a-7cdc2ed70737 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.858019] env[62600]: DEBUG nova.compute.provider_tree [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1114.015289] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223071, 'name': Rename_Task, 'duration_secs': 0.131054} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.015621] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1114.015948] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e6e5f104-d482-49ab-8967-bb74c33c95ba {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.021988] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1114.021988] env[62600]: value = "task-1223072" [ 1114.021988] env[62600]: _type = "Task" [ 1114.021988] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.029105] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223072, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.123079] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 107ed783-10a3-4f07-a809-a5a076c2b904] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1114.361536] env[62600]: DEBUG nova.scheduler.client.report [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1114.531842] env[62600]: DEBUG oslo_vmware.api [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223072, 'name': PowerOnVM_Task, 'duration_secs': 0.436016} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.532072] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1114.532288] env[62600]: INFO nova.compute.manager [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Took 6.69 seconds to spawn the instance on the hypervisor. [ 1114.532474] env[62600]: DEBUG nova.compute.manager [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1114.533280] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a443dc7b-e4c6-4749-abf2-3257431cbbc8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.626469] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: f6957b5a-5da8-4205-92e5-2f08bf948e88] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1114.867629] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.633s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.887830] env[62600]: INFO nova.scheduler.client.report [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Deleted allocations for instance eb3c5624-0599-4e81-b056-bedf34329646 [ 1115.048846] env[62600]: INFO nova.compute.manager [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Took 11.40 seconds to build instance. [ 1115.129090] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: ab11cdd1-80a7-43c4-afac-24c7930238c1] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1115.395717] env[62600]: DEBUG oslo_concurrency.lockutils [None req-a9c02572-06ab-4e09-862b-78c5dc92181e tempest-AttachInterfacesTestJSON-1224456710 tempest-AttachInterfacesTestJSON-1224456710-project-member] Lock "eb3c5624-0599-4e81-b056-bedf34329646" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.016s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.551218] env[62600]: DEBUG oslo_concurrency.lockutils [None req-84baab77-eeae-4849-af43-0bb91f028c76 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "80a9de40-675d-456b-a91b-432d2315144b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.905s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.631982] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: c82afc86-0a5b-45d7-ad7f-9a8ab9aea16f] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1115.845534] env[62600]: DEBUG nova.compute.manager [req-03f568ad-2cb0-4fbb-a7e8-59beebbcd50c req-222c3cdc-9d12-4376-baff-0f0c979b05f4 service nova] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Received event network-changed-4cf97208-acde-4474-94d5-27c1ed6ddd8c {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1115.845773] env[62600]: DEBUG nova.compute.manager [req-03f568ad-2cb0-4fbb-a7e8-59beebbcd50c req-222c3cdc-9d12-4376-baff-0f0c979b05f4 service nova] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Refreshing instance network info cache due to event network-changed-4cf97208-acde-4474-94d5-27c1ed6ddd8c. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1115.845990] env[62600]: DEBUG oslo_concurrency.lockutils [req-03f568ad-2cb0-4fbb-a7e8-59beebbcd50c req-222c3cdc-9d12-4376-baff-0f0c979b05f4 service nova] Acquiring lock "refresh_cache-80a9de40-675d-456b-a91b-432d2315144b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.846915] env[62600]: DEBUG oslo_concurrency.lockutils [req-03f568ad-2cb0-4fbb-a7e8-59beebbcd50c req-222c3cdc-9d12-4376-baff-0f0c979b05f4 service nova] Acquired lock "refresh_cache-80a9de40-675d-456b-a91b-432d2315144b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.847135] env[62600]: DEBUG nova.network.neutron [req-03f568ad-2cb0-4fbb-a7e8-59beebbcd50c req-222c3cdc-9d12-4376-baff-0f0c979b05f4 service nova] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Refreshing network info cache for port 4cf97208-acde-4474-94d5-27c1ed6ddd8c {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1116.135329] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 1b856aff-66be-4fa5-b8b7-124ac24fb4d1] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1116.638351] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: ddc80377-a14e-48d0-ad24-cc6daf7f6850] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1116.707059] env[62600]: DEBUG nova.network.neutron [req-03f568ad-2cb0-4fbb-a7e8-59beebbcd50c req-222c3cdc-9d12-4376-baff-0f0c979b05f4 service nova] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Updated VIF entry in instance network info cache for port 4cf97208-acde-4474-94d5-27c1ed6ddd8c. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1116.707487] env[62600]: DEBUG nova.network.neutron [req-03f568ad-2cb0-4fbb-a7e8-59beebbcd50c req-222c3cdc-9d12-4376-baff-0f0c979b05f4 service nova] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Updating instance_info_cache with network_info: [{"id": "4cf97208-acde-4474-94d5-27c1ed6ddd8c", "address": "fa:16:3e:eb:1c:12", "network": {"id": "e7547ac4-24a6-4778-a17a-40e79e7a0d0c", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-320671681-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef3bb131f3994f7a8673dafff6d353c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaf3dfa2-fa01-4d4d-8ecd-a9bc74d90ec2", "external-id": "nsx-vlan-transportzone-546", "segmentation_id": 546, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4cf97208-ac", "ovs_interfaceid": "4cf97208-acde-4474-94d5-27c1ed6ddd8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.142386] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: d66c404d-8fd1-4fb7-a3b9-f21854c7e735] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1117.209830] env[62600]: DEBUG oslo_concurrency.lockutils [req-03f568ad-2cb0-4fbb-a7e8-59beebbcd50c req-222c3cdc-9d12-4376-baff-0f0c979b05f4 service nova] Releasing lock "refresh_cache-80a9de40-675d-456b-a91b-432d2315144b" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.646502] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: b967fb11-e70e-4e17-b769-38da581bd83b] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1118.149214] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: ee7175b9-dbe6-4f90-bd2b-8829194dc6c3] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1118.653104] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 9d899d96-9f4f-41d1-a368-3fde5efc110a] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1118.726041] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.726041] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.726197] env[62600]: DEBUG nova.objects.instance [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lazy-loading 'pci_requests' on Instance uuid 692fa608-39f4-4d7d-9f30-b6559878b970 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1119.157066] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: c2c20e5f-6c0e-4989-bc36-ff4a27d5c534] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1119.229799] env[62600]: DEBUG nova.objects.instance [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lazy-loading 'numa_topology' on Instance uuid 692fa608-39f4-4d7d-9f30-b6559878b970 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1119.659463] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: fb6f8e36-8d24-45ea-a6e4-4d768c3b232d] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1119.732427] env[62600]: INFO nova.compute.claims [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1120.163301] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: dde6df10-618a-40a8-b33f-efc0ca3a9287] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1120.667056] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 4e1a376f-6619-4c35-b75f-b45db75815cc] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1120.784257] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48dc2a92-e33d-44ef-bdd9-7b4b65353441 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.791911] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e5789f7-34e5-4665-b41d-906570231150 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.820878] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e853298-a25a-42a5-8f00-13f84b0f7ab4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.828145] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32eb83d7-8904-4014-a18a-ea3b70a376f3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.840542] env[62600]: DEBUG nova.compute.provider_tree [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1121.170453] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: bd9b7351-01a2-429d-a860-d9ff6855eefc] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1121.343535] env[62600]: DEBUG nova.scheduler.client.report [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1121.673843] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 8f3e87a5-bb70-4e50-9fea-6bf53946b8dd] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1121.848658] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.123s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.876557] env[62600]: INFO nova.network.neutron [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Updating port 280baabd-3083-4f33-b294-1f988193665b with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1122.176942] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: afe838cc-e086-4986-87ec-4e1266bcaf60] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1122.680289] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 58bf5b37-05dd-478f-8c0b-5475ab63bf71] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1123.183656] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 734e8a2b-faaa-4165-b0e9-69eeb3699b2d] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1123.258954] env[62600]: DEBUG nova.compute.manager [req-5781cee0-4c31-4ad9-aa0b-51b90da19c61 req-323facb1-d7c9-4cb7-9fd7-407726b0bba9 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Received event network-vif-plugged-280baabd-3083-4f33-b294-1f988193665b {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1123.259176] env[62600]: DEBUG oslo_concurrency.lockutils [req-5781cee0-4c31-4ad9-aa0b-51b90da19c61 req-323facb1-d7c9-4cb7-9fd7-407726b0bba9 service nova] Acquiring lock "692fa608-39f4-4d7d-9f30-b6559878b970-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.259395] env[62600]: DEBUG oslo_concurrency.lockutils [req-5781cee0-4c31-4ad9-aa0b-51b90da19c61 req-323facb1-d7c9-4cb7-9fd7-407726b0bba9 service nova] Lock "692fa608-39f4-4d7d-9f30-b6559878b970-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.259572] env[62600]: DEBUG oslo_concurrency.lockutils [req-5781cee0-4c31-4ad9-aa0b-51b90da19c61 req-323facb1-d7c9-4cb7-9fd7-407726b0bba9 service nova] Lock "692fa608-39f4-4d7d-9f30-b6559878b970-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.259753] env[62600]: DEBUG nova.compute.manager [req-5781cee0-4c31-4ad9-aa0b-51b90da19c61 req-323facb1-d7c9-4cb7-9fd7-407726b0bba9 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] No waiting events found dispatching network-vif-plugged-280baabd-3083-4f33-b294-1f988193665b {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1123.259914] env[62600]: WARNING nova.compute.manager [req-5781cee0-4c31-4ad9-aa0b-51b90da19c61 req-323facb1-d7c9-4cb7-9fd7-407726b0bba9 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Received unexpected event network-vif-plugged-280baabd-3083-4f33-b294-1f988193665b for instance with vm_state shelved_offloaded and task_state spawning. [ 1123.345023] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1123.345194] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquired lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.345374] env[62600]: DEBUG nova.network.neutron [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1123.687021] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 92bd8b10-2034-49d4-bf38-0edc7e1e6f7d] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1124.039398] env[62600]: DEBUG nova.network.neutron [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Updating instance_info_cache with network_info: [{"id": "280baabd-3083-4f33-b294-1f988193665b", "address": "fa:16:3e:72:db:a2", "network": {"id": "dc21424f-b891-428d-a74e-76ab78d757a8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1428037886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "952323d62f034126a9b46750d03fe6ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap280baabd-30", "ovs_interfaceid": "280baabd-3083-4f33-b294-1f988193665b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.190386] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: d34c098d-a827-4dbd-96e4-e27d9d56b847] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1124.541774] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Releasing lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.566770] env[62600]: DEBUG nova.virt.hardware [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='5f573eeaf094eea12e4130bb951ddac8',container_format='bare',created_at=2024-10-09T12:08:59Z,direct_url=,disk_format='vmdk',id=10f7ff7d-b198-418a-8817-0b212a53d0a9,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-634547346-shelved',owner='952323d62f034126a9b46750d03fe6ea',properties=ImageMetaProps,protected=,size=31669248,status='active',tags=,updated_at=2024-10-09T12:09:11Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1124.567036] env[62600]: DEBUG nova.virt.hardware [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1124.567208] env[62600]: DEBUG nova.virt.hardware [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1124.567398] env[62600]: DEBUG nova.virt.hardware [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1124.567550] env[62600]: DEBUG nova.virt.hardware [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1124.567700] env[62600]: DEBUG nova.virt.hardware [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1124.567908] env[62600]: DEBUG nova.virt.hardware [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1124.568088] env[62600]: DEBUG nova.virt.hardware [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1124.568357] env[62600]: DEBUG nova.virt.hardware [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1124.568535] env[62600]: DEBUG nova.virt.hardware [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1124.568713] env[62600]: DEBUG nova.virt.hardware [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1124.569579] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc661bf-91dd-4fb1-8be2-acc3296fa05c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.577425] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b640c5a2-67af-4567-8bdf-0c4ac8b7d6b5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.590442] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:72:db:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '07e9bef1-2b0e-4e4d-997f-de71bb0e213a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '280baabd-3083-4f33-b294-1f988193665b', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1124.597744] env[62600]: DEBUG oslo.service.loopingcall [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1124.597961] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1124.598194] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6814d269-8bc2-43f3-9a03-423bb3b60b5b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.616614] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1124.616614] env[62600]: value = "task-1223074" [ 1124.616614] env[62600]: _type = "Task" [ 1124.616614] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.624478] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1223074, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.693268] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: b1cf7dc7-16bb-4e56-8955-e1bb4be6a9dc] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1125.126125] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1223074, 'name': CreateVM_Task, 'duration_secs': 0.284368} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.126284] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1125.126959] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/10f7ff7d-b198-418a-8817-0b212a53d0a9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.127157] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquired lock "[datastore1] devstack-image-cache_base/10f7ff7d-b198-418a-8817-0b212a53d0a9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.127543] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/10f7ff7d-b198-418a-8817-0b212a53d0a9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1125.127800] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d7189bc-f001-4f4d-bff2-dfdc1e92ad82 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.132168] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1125.132168] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52873450-8cbc-1732-c41a-04b29ae40474" [ 1125.132168] env[62600]: _type = "Task" [ 1125.132168] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.139612] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52873450-8cbc-1732-c41a-04b29ae40474, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.197177] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: dcbe8e2d-6ae9-465f-8394-3978ee61a15b] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1125.284696] env[62600]: DEBUG nova.compute.manager [req-21605c23-c54b-487f-a2ff-089fa437c66c req-d719657d-d167-428b-93a8-5dbe01aeb7b3 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Received event network-changed-280baabd-3083-4f33-b294-1f988193665b {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1125.284933] env[62600]: DEBUG nova.compute.manager [req-21605c23-c54b-487f-a2ff-089fa437c66c req-d719657d-d167-428b-93a8-5dbe01aeb7b3 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Refreshing instance network info cache due to event network-changed-280baabd-3083-4f33-b294-1f988193665b. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1125.285132] env[62600]: DEBUG oslo_concurrency.lockutils [req-21605c23-c54b-487f-a2ff-089fa437c66c req-d719657d-d167-428b-93a8-5dbe01aeb7b3 service nova] Acquiring lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.285287] env[62600]: DEBUG oslo_concurrency.lockutils [req-21605c23-c54b-487f-a2ff-089fa437c66c req-d719657d-d167-428b-93a8-5dbe01aeb7b3 service nova] Acquired lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.285456] env[62600]: DEBUG nova.network.neutron [req-21605c23-c54b-487f-a2ff-089fa437c66c req-d719657d-d167-428b-93a8-5dbe01aeb7b3 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Refreshing network info cache for port 280baabd-3083-4f33-b294-1f988193665b {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1125.642792] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Releasing lock "[datastore1] devstack-image-cache_base/10f7ff7d-b198-418a-8817-0b212a53d0a9" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.643078] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Processing image 10f7ff7d-b198-418a-8817-0b212a53d0a9 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1125.643319] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/10f7ff7d-b198-418a-8817-0b212a53d0a9/10f7ff7d-b198-418a-8817-0b212a53d0a9.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.643474] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquired lock "[datastore1] devstack-image-cache_base/10f7ff7d-b198-418a-8817-0b212a53d0a9/10f7ff7d-b198-418a-8817-0b212a53d0a9.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.643654] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1125.643897] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ac4da33a-71fd-45bc-b502-6c8dd2298076 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.651441] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1125.651613] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1125.652290] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1029c50-65fc-46b1-b279-bab767fc5b03 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.656948] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1125.656948] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520c6f0a-a262-6e0d-bc79-8dc79c07d021" [ 1125.656948] env[62600]: _type = "Task" [ 1125.656948] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.663735] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520c6f0a-a262-6e0d-bc79-8dc79c07d021, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.700270] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 0a8840f1-7681-4aaa-9f3a-32b72c04c1c4] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1125.987343] env[62600]: DEBUG nova.network.neutron [req-21605c23-c54b-487f-a2ff-089fa437c66c req-d719657d-d167-428b-93a8-5dbe01aeb7b3 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Updated VIF entry in instance network info cache for port 280baabd-3083-4f33-b294-1f988193665b. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1125.987708] env[62600]: DEBUG nova.network.neutron [req-21605c23-c54b-487f-a2ff-089fa437c66c req-d719657d-d167-428b-93a8-5dbe01aeb7b3 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Updating instance_info_cache with network_info: [{"id": "280baabd-3083-4f33-b294-1f988193665b", "address": "fa:16:3e:72:db:a2", "network": {"id": "dc21424f-b891-428d-a74e-76ab78d757a8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1428037886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "952323d62f034126a9b46750d03fe6ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap280baabd-30", "ovs_interfaceid": "280baabd-3083-4f33-b294-1f988193665b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1126.166305] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Preparing fetch location {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1126.166568] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Fetch image to [datastore1] OSTACK_IMG_4559fe29-7b35-4383-b63b-0ba85bb96da4/OSTACK_IMG_4559fe29-7b35-4383-b63b-0ba85bb96da4.vmdk {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1126.166764] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Downloading stream optimized image 10f7ff7d-b198-418a-8817-0b212a53d0a9 to [datastore1] OSTACK_IMG_4559fe29-7b35-4383-b63b-0ba85bb96da4/OSTACK_IMG_4559fe29-7b35-4383-b63b-0ba85bb96da4.vmdk on the data store datastore1 as vApp {{(pid=62600) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1126.166942] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Downloading image file data 10f7ff7d-b198-418a-8817-0b212a53d0a9 to the ESX as VM named 'OSTACK_IMG_4559fe29-7b35-4383-b63b-0ba85bb96da4' {{(pid=62600) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1126.203519] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: e25c631d-3d1b-40d2-9fb1-a65431f991a9] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1126.241437] env[62600]: DEBUG oslo_vmware.rw_handles [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1126.241437] env[62600]: value = "resgroup-9" [ 1126.241437] env[62600]: _type = "ResourcePool" [ 1126.241437] env[62600]: }. {{(pid=62600) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1126.241985] env[62600]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-7b53185e-c839-4cde-82aa-9df888be0160 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.262846] env[62600]: DEBUG oslo_vmware.rw_handles [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lease: (returnval){ [ 1126.262846] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ff7080-3e5f-786f-0b2a-8552e0cec9b1" [ 1126.262846] env[62600]: _type = "HttpNfcLease" [ 1126.262846] env[62600]: } obtained for vApp import into resource pool (val){ [ 1126.262846] env[62600]: value = "resgroup-9" [ 1126.262846] env[62600]: _type = "ResourcePool" [ 1126.262846] env[62600]: }. {{(pid=62600) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1126.263155] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the lease: (returnval){ [ 1126.263155] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ff7080-3e5f-786f-0b2a-8552e0cec9b1" [ 1126.263155] env[62600]: _type = "HttpNfcLease" [ 1126.263155] env[62600]: } to be ready. {{(pid=62600) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1126.268967] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1126.268967] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ff7080-3e5f-786f-0b2a-8552e0cec9b1" [ 1126.268967] env[62600]: _type = "HttpNfcLease" [ 1126.268967] env[62600]: } is initializing. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1126.490134] env[62600]: DEBUG oslo_concurrency.lockutils [req-21605c23-c54b-487f-a2ff-089fa437c66c req-d719657d-d167-428b-93a8-5dbe01aeb7b3 service nova] Releasing lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.707285] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: a358a3cb-deda-419a-aa3c-ce7aeb534240] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1126.773071] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1126.773071] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ff7080-3e5f-786f-0b2a-8552e0cec9b1" [ 1126.773071] env[62600]: _type = "HttpNfcLease" [ 1126.773071] env[62600]: } is ready. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1126.773361] env[62600]: DEBUG oslo_vmware.rw_handles [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1126.773361] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ff7080-3e5f-786f-0b2a-8552e0cec9b1" [ 1126.773361] env[62600]: _type = "HttpNfcLease" [ 1126.773361] env[62600]: }. {{(pid=62600) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1126.774019] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d92d43-1d51-4efb-a10d-b86dfbfa1aea {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.780528] env[62600]: DEBUG oslo_vmware.rw_handles [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527856e2-11c9-9b69-54e7-a32b016a5ca8/disk-0.vmdk from lease info. {{(pid=62600) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1126.780709] env[62600]: DEBUG oslo_vmware.rw_handles [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Creating HTTP connection to write to file with size = 31669248 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527856e2-11c9-9b69-54e7-a32b016a5ca8/disk-0.vmdk. {{(pid=62600) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1126.842250] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1496d5b2-40ba-4741-8d09-d2eed0300cf7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.210378] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: efff4d72-4673-4a8d-9a81-be9d700ff881] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1127.713443] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 92cf1acb-b471-4add-a73f-7e9e94fbaaa4] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1127.871039] env[62600]: DEBUG oslo_vmware.rw_handles [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Completed reading data from the image iterator. {{(pid=62600) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1127.871315] env[62600]: DEBUG oslo_vmware.rw_handles [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527856e2-11c9-9b69-54e7-a32b016a5ca8/disk-0.vmdk. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1127.872349] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed346d29-0946-4bae-a5b4-369093fac93c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.879733] env[62600]: DEBUG oslo_vmware.rw_handles [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527856e2-11c9-9b69-54e7-a32b016a5ca8/disk-0.vmdk is in state: ready. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1127.879899] env[62600]: DEBUG oslo_vmware.rw_handles [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527856e2-11c9-9b69-54e7-a32b016a5ca8/disk-0.vmdk. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1127.880149] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-ec0b4d15-42eb-4848-a249-d9e811b1866d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.070732] env[62600]: DEBUG oslo_vmware.rw_handles [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527856e2-11c9-9b69-54e7-a32b016a5ca8/disk-0.vmdk. {{(pid=62600) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1128.070961] env[62600]: INFO nova.virt.vmwareapi.images [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Downloaded image file data 10f7ff7d-b198-418a-8817-0b212a53d0a9 [ 1128.071794] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff6682d-fe6c-4b84-94bf-6bf819e974ba {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.087986] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bf4ae2e7-f576-4c8c-9d67-36b78e7aa09f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.134619] env[62600]: INFO nova.virt.vmwareapi.images [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] The imported VM was unregistered [ 1128.137135] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Caching image {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1128.137375] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Creating directory with path [datastore1] devstack-image-cache_base/10f7ff7d-b198-418a-8817-0b212a53d0a9 {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1128.137646] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d46ee07f-7f87-40cc-81d8-c6c98f0a2e8d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.216427] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: de4f9304-3357-4eaa-9c94-fe28bc554086] Instance has had 0 of 5 cleanup attempts {{(pid=62600) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1128.220796] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Created directory with path [datastore1] devstack-image-cache_base/10f7ff7d-b198-418a-8817-0b212a53d0a9 {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1128.220997] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_4559fe29-7b35-4383-b63b-0ba85bb96da4/OSTACK_IMG_4559fe29-7b35-4383-b63b-0ba85bb96da4.vmdk to [datastore1] devstack-image-cache_base/10f7ff7d-b198-418a-8817-0b212a53d0a9/10f7ff7d-b198-418a-8817-0b212a53d0a9.vmdk. {{(pid=62600) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1128.221275] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-11f9e324-716f-4c06-baa9-694b0ba9db50 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.227870] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1128.227870] env[62600]: value = "task-1223077" [ 1128.227870] env[62600]: _type = "Task" [ 1128.227870] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.235782] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223077, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.720518] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1128.720725] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Cleaning up deleted instances with incomplete migration {{(pid=62600) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1128.740616] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223077, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.224207] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1129.241254] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223077, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.740541] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223077, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.243630] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223077, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.726972] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1130.727208] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1130.741097] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223077, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.318417} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.741978] env[62600]: INFO nova.virt.vmwareapi.ds_util [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_4559fe29-7b35-4383-b63b-0ba85bb96da4/OSTACK_IMG_4559fe29-7b35-4383-b63b-0ba85bb96da4.vmdk to [datastore1] devstack-image-cache_base/10f7ff7d-b198-418a-8817-0b212a53d0a9/10f7ff7d-b198-418a-8817-0b212a53d0a9.vmdk. [ 1130.742195] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Cleaning up location [datastore1] OSTACK_IMG_4559fe29-7b35-4383-b63b-0ba85bb96da4 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1130.742365] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_4559fe29-7b35-4383-b63b-0ba85bb96da4 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1130.742612] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1f303ec4-a471-435e-bd2b-69c86fcba625 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.748421] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1130.748421] env[62600]: value = "task-1223078" [ 1130.748421] env[62600]: _type = "Task" [ 1130.748421] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.755669] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223078, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.231892] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1131.232111] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Starting heal instance info cache {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1131.232214] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Rebuilding the list of instances to heal {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1131.258071] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223078, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.034307} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.258392] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1131.258490] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Releasing lock "[datastore1] devstack-image-cache_base/10f7ff7d-b198-418a-8817-0b212a53d0a9/10f7ff7d-b198-418a-8817-0b212a53d0a9.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.258731] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/10f7ff7d-b198-418a-8817-0b212a53d0a9/10f7ff7d-b198-418a-8817-0b212a53d0a9.vmdk to [datastore1] 692fa608-39f4-4d7d-9f30-b6559878b970/692fa608-39f4-4d7d-9f30-b6559878b970.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1131.258975] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d3d1978-36bd-4cff-88e8-3cc5bf0d6263 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.266306] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1131.266306] env[62600]: value = "task-1223079" [ 1131.266306] env[62600]: _type = "Task" [ 1131.266306] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.273149] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223079, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.764521] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "refresh_cache-c2f38d56-72fa-4bf0-a2cb-13d666571466" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.764686] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquired lock "refresh_cache-c2f38d56-72fa-4bf0-a2cb-13d666571466" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.764860] env[62600]: DEBUG nova.network.neutron [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Forcefully refreshing network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1131.765036] env[62600]: DEBUG nova.objects.instance [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lazy-loading 'info_cache' on Instance uuid c2f38d56-72fa-4bf0-a2cb-13d666571466 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1131.777067] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223079, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.285514] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223079, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.782576] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223079, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.284982] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223079, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.514180] env[62600]: DEBUG nova.network.neutron [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Updating instance_info_cache with network_info: [{"id": "c715a88c-6498-428e-989d-3a1aa2f2e3bd", "address": "fa:16:3e:67:c1:e4", "network": {"id": "7ed8725a-d4a1-4d66-8bd3-08c9abf86064", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1314620504-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4e088a25c01c40d98d27e22563b4c625", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b98c49ac-0eb7-4311-aa8f-60581b2ce706", "external-id": "nsx-vlan-transportzone-184", "segmentation_id": 184, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc715a88c-64", "ovs_interfaceid": "c715a88c-6498-428e-989d-3a1aa2f2e3bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.783015] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223079, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.344099} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.785211] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/10f7ff7d-b198-418a-8817-0b212a53d0a9/10f7ff7d-b198-418a-8817-0b212a53d0a9.vmdk to [datastore1] 692fa608-39f4-4d7d-9f30-b6559878b970/692fa608-39f4-4d7d-9f30-b6559878b970.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1133.785211] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f91e2792-eb9c-441e-83a3-1a5dd46efa7c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.806396] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 692fa608-39f4-4d7d-9f30-b6559878b970/692fa608-39f4-4d7d-9f30-b6559878b970.vmdk or device None with type streamOptimized {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1133.806639] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c336beb1-7c88-430e-8219-2cc666df4004 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.825356] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1133.825356] env[62600]: value = "task-1223080" [ 1133.825356] env[62600]: _type = "Task" [ 1133.825356] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.832517] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223080, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.016650] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Releasing lock "refresh_cache-c2f38d56-72fa-4bf0-a2cb-13d666571466" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.016867] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Updated the network info_cache for instance {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1134.017095] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.017264] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.017410] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.017557] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.017700] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._sync_power_states {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.335344] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223080, 'name': ReconfigVM_Task, 'duration_secs': 0.273395} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.335625] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 692fa608-39f4-4d7d-9f30-b6559878b970/692fa608-39f4-4d7d-9f30-b6559878b970.vmdk or device None with type streamOptimized {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1134.337106] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'boot_index': 0, 'device_name': '/dev/sda', 'encrypted': False, 'guest_format': None, 'encryption_secret_uuid': None, 'size': 0, 'device_type': 'disk', 'encryption_options': None, 'disk_bus': None, 'encryption_format': None, 'image_id': 'e7bd1d77-bfff-4684-9545-f6a1c69efe58'}], 'ephemerals': [], 'block_device_mapping': [{'boot_index': None, 'mount_device': '/dev/sdb', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264377', 'volume_id': '34095f5d-35c4-4bad-88a3-e3e627c33b5c', 'name': 'volume-34095f5d-35c4-4bad-88a3-e3e627c33b5c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '692fa608-39f4-4d7d-9f30-b6559878b970', 'attached_at': '', 'detached_at': '', 'volume_id': '34095f5d-35c4-4bad-88a3-e3e627c33b5c', 'serial': '34095f5d-35c4-4bad-88a3-e3e627c33b5c'}, 'guest_format': None, 'attachment_id': '8d50d655-28eb-47a6-93fa-14394369ca57', 'delete_on_termination': False, 'device_type': None, 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=62600) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1134.337334] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Volume attach. Driver type: vmdk {{(pid=62600) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1134.337532] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264377', 'volume_id': '34095f5d-35c4-4bad-88a3-e3e627c33b5c', 'name': 'volume-34095f5d-35c4-4bad-88a3-e3e627c33b5c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '692fa608-39f4-4d7d-9f30-b6559878b970', 'attached_at': '', 'detached_at': '', 'volume_id': '34095f5d-35c4-4bad-88a3-e3e627c33b5c', 'serial': '34095f5d-35c4-4bad-88a3-e3e627c33b5c'} {{(pid=62600) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1134.338276] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfb6b376-b21f-410e-a069-60e9a7c2bec6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.352968] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7e64055-d20b-4eec-b5ee-9cd5ceb46b99 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.375634] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] volume-34095f5d-35c4-4bad-88a3-e3e627c33b5c/volume-34095f5d-35c4-4bad-88a3-e3e627c33b5c.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1134.375896] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b5bb9d2-ccb3-49ed-b571-d1af30b088a3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.392749] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1134.392749] env[62600]: value = "task-1223081" [ 1134.392749] env[62600]: _type = "Task" [ 1134.392749] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.399792] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223081, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.521461] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Getting list of instances from cluster (obj){ [ 1134.521461] env[62600]: value = "domain-c8" [ 1134.521461] env[62600]: _type = "ClusterComputeResource" [ 1134.521461] env[62600]: } {{(pid=62600) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1134.522541] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3e2d538-1993-4a7d-8cff-067a739125cd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.534621] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Got total of 3 instances {{(pid=62600) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1134.534807] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Triggering sync for uuid c2f38d56-72fa-4bf0-a2cb-13d666571466 {{(pid=62600) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1134.535018] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Triggering sync for uuid 692fa608-39f4-4d7d-9f30-b6559878b970 {{(pid=62600) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1134.535190] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Triggering sync for uuid 80a9de40-675d-456b-a91b-432d2315144b {{(pid=62600) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1134.535510] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "c2f38d56-72fa-4bf0-a2cb-13d666571466" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.535733] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "c2f38d56-72fa-4bf0-a2cb-13d666571466" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.536016] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "692fa608-39f4-4d7d-9f30-b6559878b970" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.536250] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "80a9de40-675d-456b-a91b-432d2315144b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.536440] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "80a9de40-675d-456b-a91b-432d2315144b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.536640] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.536774] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62600) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1134.537455] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d93a789-d870-4cc1-b751-7055562578a5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.540181] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0ace2dd-f66c-461a-8c58-a5d05988134e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.902543] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223081, 'name': ReconfigVM_Task, 'duration_secs': 0.287465} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.902843] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Reconfigured VM instance instance-00000064 to attach disk [datastore1] volume-34095f5d-35c4-4bad-88a3-e3e627c33b5c/volume-34095f5d-35c4-4bad-88a3-e3e627c33b5c.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1134.907489] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2dff815c-1874-4717-9224-7d58c5810fab {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.921861] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1134.921861] env[62600]: value = "task-1223082" [ 1134.921861] env[62600]: _type = "Task" [ 1134.921861] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.929232] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223082, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.050096] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "c2f38d56-72fa-4bf0-a2cb-13d666571466" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.514s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.053681] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "80a9de40-675d-456b-a91b-432d2315144b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.517s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.431964] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223082, 'name': ReconfigVM_Task, 'duration_secs': 0.134092} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.432299] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264377', 'volume_id': '34095f5d-35c4-4bad-88a3-e3e627c33b5c', 'name': 'volume-34095f5d-35c4-4bad-88a3-e3e627c33b5c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '692fa608-39f4-4d7d-9f30-b6559878b970', 'attached_at': '', 'detached_at': '', 'volume_id': '34095f5d-35c4-4bad-88a3-e3e627c33b5c', 'serial': '34095f5d-35c4-4bad-88a3-e3e627c33b5c'} {{(pid=62600) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1135.432885] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bf00659e-05ba-4217-bd79-d90a2798aae2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.439388] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1135.439388] env[62600]: value = "task-1223083" [ 1135.439388] env[62600]: _type = "Task" [ 1135.439388] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.447951] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223083, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.948840] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223083, 'name': Rename_Task, 'duration_secs': 0.136623} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.949233] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1135.949482] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dd0bc126-1ec0-47ba-be52-236330d5a44d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.955563] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1135.955563] env[62600]: value = "task-1223084" [ 1135.955563] env[62600]: _type = "Task" [ 1135.955563] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.962451] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223084, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.465534] env[62600]: DEBUG oslo_vmware.api [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223084, 'name': PowerOnVM_Task, 'duration_secs': 0.415698} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.465760] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1136.569400] env[62600]: DEBUG nova.compute.manager [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1136.570330] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8cfa277-2b42-4628-92ae-48a58a94432c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.087012] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e68a035a-cf65-4ff3-b4e8-8f914751a907 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "692fa608-39f4-4d7d-9f30-b6559878b970" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 25.672s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.087878] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "692fa608-39f4-4d7d-9f30-b6559878b970" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 2.552s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.088150] env[62600]: INFO nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] During sync_power_state the instance has a pending task (spawning). Skip. [ 1137.088910] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "692fa608-39f4-4d7d-9f30-b6559878b970" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.770586] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "80a9de40-675d-456b-a91b-432d2315144b" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.770908] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "80a9de40-675d-456b-a91b-432d2315144b" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.273988] env[62600]: DEBUG nova.compute.utils [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1155.776907] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "80a9de40-675d-456b-a91b-432d2315144b" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1156.840683] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "80a9de40-675d-456b-a91b-432d2315144b" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.841087] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "80a9de40-675d-456b-a91b-432d2315144b" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.841260] env[62600]: INFO nova.compute.manager [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Attaching volume 26a1aca3-82f8-4ca8-8403-3c64a1246167 to /dev/sdb [ 1156.903552] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92020f87-ceb3-4d63-ac10-0ef70fb113e4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.911585] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e75b902-400a-4706-a3ba-70db23f482bf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.924109] env[62600]: DEBUG nova.virt.block_device [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Updating existing volume attachment record: 4f9b42f4-6039-4737-90cb-23e3124f4020 {{(pid=62600) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1158.252399] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1158.728059] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1159.727551] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1159.727907] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1160.231463] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.231755] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.231875] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1160.232045] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62600) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1160.232954] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2404ef06-c368-493e-9e51-2ce367789b47 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.240976] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef3b1b7e-051e-43dc-98d1-c18fe2d20d69 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.254560] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb63f9ff-45db-48bd-b8d9-3cf040b9b186 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.260424] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4bfb377-bf3c-4724-abb7-ec103ffd33c1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.288231] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181025MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=62600) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1160.288366] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.288548] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.313132] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance c2f38d56-72fa-4bf0-a2cb-13d666571466 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1161.313427] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 80a9de40-675d-456b-a91b-432d2315144b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1161.313427] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 692fa608-39f4-4d7d-9f30-b6559878b970 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1161.313587] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1161.313723] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1161.357819] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26ee6134-125a-4f68-87f2-e6ed27d16ecd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.365253] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cafd637f-2585-4634-b386-182a101d6cf6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.395530] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c84c264f-df9a-4d9b-8cb6-76b59c9e411c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.402142] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b7c658-1f0c-41f4-85c6-c33e9a22d9d5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.414704] env[62600]: DEBUG nova.compute.provider_tree [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1161.508862] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Volume attach. Driver type: vmdk {{(pid=62600) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1161.509123] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264380', 'volume_id': '26a1aca3-82f8-4ca8-8403-3c64a1246167', 'name': 'volume-26a1aca3-82f8-4ca8-8403-3c64a1246167', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '80a9de40-675d-456b-a91b-432d2315144b', 'attached_at': '', 'detached_at': '', 'volume_id': '26a1aca3-82f8-4ca8-8403-3c64a1246167', 'serial': '26a1aca3-82f8-4ca8-8403-3c64a1246167'} {{(pid=62600) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1161.509960] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e60019-6495-4407-b8dc-020e6500b667 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.525932] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff0a2649-6083-4725-89e2-773e249c5165 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.549454] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] volume-26a1aca3-82f8-4ca8-8403-3c64a1246167/volume-26a1aca3-82f8-4ca8-8403-3c64a1246167.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1161.549687] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-137c4aa3-69ff-4c78-8240-1146ec6111de {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.566949] env[62600]: DEBUG oslo_vmware.api [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1161.566949] env[62600]: value = "task-1223087" [ 1161.566949] env[62600]: _type = "Task" [ 1161.566949] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.573964] env[62600]: DEBUG oslo_vmware.api [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223087, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.917569] env[62600]: DEBUG nova.scheduler.client.report [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1162.076350] env[62600]: DEBUG oslo_vmware.api [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223087, 'name': ReconfigVM_Task, 'duration_secs': 0.309659} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.076619] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Reconfigured VM instance instance-00000068 to attach disk [datastore2] volume-26a1aca3-82f8-4ca8-8403-3c64a1246167/volume-26a1aca3-82f8-4ca8-8403-3c64a1246167.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1162.081073] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1d78f19-b9be-4955-80f2-b2642a67d477 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.094854] env[62600]: DEBUG oslo_vmware.api [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1162.094854] env[62600]: value = "task-1223088" [ 1162.094854] env[62600]: _type = "Task" [ 1162.094854] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.102148] env[62600]: DEBUG oslo_vmware.api [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223088, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.421986] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62600) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1162.422325] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.134s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.604223] env[62600]: DEBUG oslo_vmware.api [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223088, 'name': ReconfigVM_Task, 'duration_secs': 0.146791} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.604526] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264380', 'volume_id': '26a1aca3-82f8-4ca8-8403-3c64a1246167', 'name': 'volume-26a1aca3-82f8-4ca8-8403-3c64a1246167', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '80a9de40-675d-456b-a91b-432d2315144b', 'attached_at': '', 'detached_at': '', 'volume_id': '26a1aca3-82f8-4ca8-8403-3c64a1246167', 'serial': '26a1aca3-82f8-4ca8-8403-3c64a1246167'} {{(pid=62600) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1163.422491] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1163.422817] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1163.422906] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Starting heal instance info cache {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1163.641153] env[62600]: DEBUG nova.objects.instance [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lazy-loading 'flavor' on Instance uuid 80a9de40-675d-456b-a91b-432d2315144b {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1163.972943] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1163.973119] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquired lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1163.973274] env[62600]: DEBUG nova.network.neutron [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Forcefully refreshing network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1164.146838] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1fec10e4-ca56-49b6-9242-51f3aacd6afa tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "80a9de40-675d-456b-a91b-432d2315144b" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.306s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.270955] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9d1db9fa-0c8f-490c-a11c-b6ea2a7cd152 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "80a9de40-675d-456b-a91b-432d2315144b" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1164.271217] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9d1db9fa-0c8f-490c-a11c-b6ea2a7cd152 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "80a9de40-675d-456b-a91b-432d2315144b" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.774156] env[62600]: INFO nova.compute.manager [None req-9d1db9fa-0c8f-490c-a11c-b6ea2a7cd152 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Detaching volume 26a1aca3-82f8-4ca8-8403-3c64a1246167 [ 1164.803870] env[62600]: INFO nova.virt.block_device [None req-9d1db9fa-0c8f-490c-a11c-b6ea2a7cd152 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Attempting to driver detach volume 26a1aca3-82f8-4ca8-8403-3c64a1246167 from mountpoint /dev/sdb [ 1164.804131] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d1db9fa-0c8f-490c-a11c-b6ea2a7cd152 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Volume detach. Driver type: vmdk {{(pid=62600) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1164.804310] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d1db9fa-0c8f-490c-a11c-b6ea2a7cd152 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264380', 'volume_id': '26a1aca3-82f8-4ca8-8403-3c64a1246167', 'name': 'volume-26a1aca3-82f8-4ca8-8403-3c64a1246167', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '80a9de40-675d-456b-a91b-432d2315144b', 'attached_at': '', 'detached_at': '', 'volume_id': '26a1aca3-82f8-4ca8-8403-3c64a1246167', 'serial': '26a1aca3-82f8-4ca8-8403-3c64a1246167'} {{(pid=62600) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1164.805210] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a938b98-e326-493f-a485-66610e0647cf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.827946] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d867f316-e951-4146-9cf5-d4e093ffc419 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.834312] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86cb90f2-67a6-4037-92d2-7e04783024f0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.853227] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d046b9ea-f1db-4135-961e-c72336a56c4c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.867007] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d1db9fa-0c8f-490c-a11c-b6ea2a7cd152 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] The volume has not been displaced from its original location: [datastore2] volume-26a1aca3-82f8-4ca8-8403-3c64a1246167/volume-26a1aca3-82f8-4ca8-8403-3c64a1246167.vmdk. No consolidation needed. {{(pid=62600) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1164.872153] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d1db9fa-0c8f-490c-a11c-b6ea2a7cd152 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Reconfiguring VM instance instance-00000068 to detach disk 2001 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1164.872405] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-679b8f87-fb3c-4a43-8a85-1ddacea6b9e4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.889375] env[62600]: DEBUG oslo_vmware.api [None req-9d1db9fa-0c8f-490c-a11c-b6ea2a7cd152 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1164.889375] env[62600]: value = "task-1223089" [ 1164.889375] env[62600]: _type = "Task" [ 1164.889375] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.896344] env[62600]: DEBUG oslo_vmware.api [None req-9d1db9fa-0c8f-490c-a11c-b6ea2a7cd152 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223089, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.194183] env[62600]: DEBUG nova.network.neutron [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Updating instance_info_cache with network_info: [{"id": "280baabd-3083-4f33-b294-1f988193665b", "address": "fa:16:3e:72:db:a2", "network": {"id": "dc21424f-b891-428d-a74e-76ab78d757a8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1428037886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "952323d62f034126a9b46750d03fe6ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap280baabd-30", "ovs_interfaceid": "280baabd-3083-4f33-b294-1f988193665b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.399487] env[62600]: DEBUG oslo_vmware.api [None req-9d1db9fa-0c8f-490c-a11c-b6ea2a7cd152 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223089, 'name': ReconfigVM_Task, 'duration_secs': 0.212386} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.399765] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d1db9fa-0c8f-490c-a11c-b6ea2a7cd152 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Reconfigured VM instance instance-00000068 to detach disk 2001 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1165.404363] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e4c9baa-4d17-4e06-903e-b896ec708afa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.419201] env[62600]: DEBUG oslo_vmware.api [None req-9d1db9fa-0c8f-490c-a11c-b6ea2a7cd152 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1165.419201] env[62600]: value = "task-1223090" [ 1165.419201] env[62600]: _type = "Task" [ 1165.419201] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.427037] env[62600]: DEBUG oslo_vmware.api [None req-9d1db9fa-0c8f-490c-a11c-b6ea2a7cd152 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223090, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.696955] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Releasing lock "refresh_cache-692fa608-39f4-4d7d-9f30-b6559878b970" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1165.697202] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Updated the network info_cache for instance {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1165.697410] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1165.697574] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1165.727701] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1165.727855] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62600) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1165.929024] env[62600]: DEBUG oslo_vmware.api [None req-9d1db9fa-0c8f-490c-a11c-b6ea2a7cd152 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223090, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.430037] env[62600]: DEBUG oslo_vmware.api [None req-9d1db9fa-0c8f-490c-a11c-b6ea2a7cd152 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223090, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.930729] env[62600]: DEBUG oslo_vmware.api [None req-9d1db9fa-0c8f-490c-a11c-b6ea2a7cd152 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223090, 'name': ReconfigVM_Task, 'duration_secs': 1.125895} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.931064] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d1db9fa-0c8f-490c-a11c-b6ea2a7cd152 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264380', 'volume_id': '26a1aca3-82f8-4ca8-8403-3c64a1246167', 'name': 'volume-26a1aca3-82f8-4ca8-8403-3c64a1246167', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '80a9de40-675d-456b-a91b-432d2315144b', 'attached_at': '', 'detached_at': '', 'volume_id': '26a1aca3-82f8-4ca8-8403-3c64a1246167', 'serial': '26a1aca3-82f8-4ca8-8403-3c64a1246167'} {{(pid=62600) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1166.962695] env[62600]: DEBUG oslo_concurrency.lockutils [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Acquiring lock "c2f38d56-72fa-4bf0-a2cb-13d666571466" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.963048] env[62600]: DEBUG oslo_concurrency.lockutils [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Lock "c2f38d56-72fa-4bf0-a2cb-13d666571466" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.963278] env[62600]: DEBUG oslo_concurrency.lockutils [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Acquiring lock "c2f38d56-72fa-4bf0-a2cb-13d666571466-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.963476] env[62600]: DEBUG oslo_concurrency.lockutils [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Lock "c2f38d56-72fa-4bf0-a2cb-13d666571466-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.963649] env[62600]: DEBUG oslo_concurrency.lockutils [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Lock "c2f38d56-72fa-4bf0-a2cb-13d666571466-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1166.965696] env[62600]: INFO nova.compute.manager [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Terminating instance [ 1166.967422] env[62600]: DEBUG nova.compute.manager [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1166.967620] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1166.968432] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e1c21c-53fe-48c7-9db3-1fc6a16d0432 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.977549] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1166.977779] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6d9fdad1-5c86-45de-b41d-272dcc9e24dc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.983814] env[62600]: DEBUG oslo_vmware.api [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Waiting for the task: (returnval){ [ 1166.983814] env[62600]: value = "task-1223091" [ 1166.983814] env[62600]: _type = "Task" [ 1166.983814] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.991342] env[62600]: DEBUG oslo_vmware.api [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': task-1223091, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.473100] env[62600]: DEBUG nova.objects.instance [None req-9d1db9fa-0c8f-490c-a11c-b6ea2a7cd152 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lazy-loading 'flavor' on Instance uuid 80a9de40-675d-456b-a91b-432d2315144b {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1167.493585] env[62600]: DEBUG oslo_vmware.api [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': task-1223091, 'name': PowerOffVM_Task, 'duration_secs': 0.20241} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.493845] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1167.494032] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1167.494282] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3cf2d203-f8ad-43f9-8033-5e85403783e3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.555433] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1167.555665] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1167.555856] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Deleting the datastore file [datastore2] c2f38d56-72fa-4bf0-a2cb-13d666571466 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1167.556120] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96695916-f14e-4be6-ac9c-39e578a3de9c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.562667] env[62600]: DEBUG oslo_vmware.api [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Waiting for the task: (returnval){ [ 1167.562667] env[62600]: value = "task-1223093" [ 1167.562667] env[62600]: _type = "Task" [ 1167.562667] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.569678] env[62600]: DEBUG oslo_vmware.api [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': task-1223093, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.072075] env[62600]: DEBUG oslo_vmware.api [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Task: {'id': task-1223093, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127489} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.072464] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1168.072593] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1168.072843] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1168.073025] env[62600]: INFO nova.compute.manager [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1168.073284] env[62600]: DEBUG oslo.service.loopingcall [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1168.073481] env[62600]: DEBUG nova.compute.manager [-] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1168.073577] env[62600]: DEBUG nova.network.neutron [-] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1168.480081] env[62600]: DEBUG oslo_concurrency.lockutils [None req-9d1db9fa-0c8f-490c-a11c-b6ea2a7cd152 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "80a9de40-675d-456b-a91b-432d2315144b" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.209s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.553191] env[62600]: DEBUG nova.compute.manager [req-8c7cf954-b7a0-47dd-8c34-e459bb45ad30 req-c661967e-fe54-417f-a3b0-d7dbae4deffb service nova] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Received event network-vif-deleted-c715a88c-6498-428e-989d-3a1aa2f2e3bd {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1168.553191] env[62600]: INFO nova.compute.manager [req-8c7cf954-b7a0-47dd-8c34-e459bb45ad30 req-c661967e-fe54-417f-a3b0-d7dbae4deffb service nova] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Neutron deleted interface c715a88c-6498-428e-989d-3a1aa2f2e3bd; detaching it from the instance and deleting it from the info cache [ 1168.553191] env[62600]: DEBUG nova.network.neutron [req-8c7cf954-b7a0-47dd-8c34-e459bb45ad30 req-c661967e-fe54-417f-a3b0-d7dbae4deffb service nova] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1168.994236] env[62600]: DEBUG nova.network.neutron [-] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1169.055523] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f0a12f2-6dfc-419b-bd49-ec3c988ec462 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.066491] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c0c98ec-0605-480c-a4d0-6e6402bac715 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.093176] env[62600]: DEBUG nova.compute.manager [req-8c7cf954-b7a0-47dd-8c34-e459bb45ad30 req-c661967e-fe54-417f-a3b0-d7dbae4deffb service nova] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Detach interface failed, port_id=c715a88c-6498-428e-989d-3a1aa2f2e3bd, reason: Instance c2f38d56-72fa-4bf0-a2cb-13d666571466 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1169.497221] env[62600]: INFO nova.compute.manager [-] [instance: c2f38d56-72fa-4bf0-a2cb-13d666571466] Took 1.42 seconds to deallocate network for instance. [ 1169.551076] env[62600]: DEBUG oslo_concurrency.lockutils [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "80a9de40-675d-456b-a91b-432d2315144b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.551415] env[62600]: DEBUG oslo_concurrency.lockutils [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "80a9de40-675d-456b-a91b-432d2315144b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.552180] env[62600]: DEBUG oslo_concurrency.lockutils [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "80a9de40-675d-456b-a91b-432d2315144b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.552477] env[62600]: DEBUG oslo_concurrency.lockutils [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "80a9de40-675d-456b-a91b-432d2315144b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.552700] env[62600]: DEBUG oslo_concurrency.lockutils [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "80a9de40-675d-456b-a91b-432d2315144b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.554690] env[62600]: INFO nova.compute.manager [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Terminating instance [ 1169.556544] env[62600]: DEBUG nova.compute.manager [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1169.556792] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1169.557627] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20a45178-c93e-4b14-b695-78b578456bec {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.565166] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1169.565389] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2bfa3c3f-1726-4c3a-acf7-91ee403c7420 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.571848] env[62600]: DEBUG oslo_vmware.api [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1169.571848] env[62600]: value = "task-1223094" [ 1169.571848] env[62600]: _type = "Task" [ 1169.571848] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.579542] env[62600]: DEBUG oslo_vmware.api [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223094, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.003947] env[62600]: DEBUG oslo_concurrency.lockutils [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.004256] env[62600]: DEBUG oslo_concurrency.lockutils [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.004483] env[62600]: DEBUG nova.objects.instance [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Lazy-loading 'resources' on Instance uuid c2f38d56-72fa-4bf0-a2cb-13d666571466 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1170.082196] env[62600]: DEBUG oslo_vmware.api [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223094, 'name': PowerOffVM_Task, 'duration_secs': 0.194838} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.082499] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1170.082677] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1170.082950] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ed8b8f7a-8446-427f-b556-a7e0e278af26 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.147504] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1170.147891] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1170.147936] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Deleting the datastore file [datastore2] 80a9de40-675d-456b-a91b-432d2315144b {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1170.148195] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-58ca3e6c-0f92-46c8-b053-92e1109a115b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.154249] env[62600]: DEBUG oslo_vmware.api [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for the task: (returnval){ [ 1170.154249] env[62600]: value = "task-1223096" [ 1170.154249] env[62600]: _type = "Task" [ 1170.154249] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.161548] env[62600]: DEBUG oslo_vmware.api [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223096, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.555413] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-747b9c98-4d58-42fe-9dd5-897a306811bc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.562930] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f83c306-473d-43d6-bbfc-0b7764b7521a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.591488] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d9a16a-26d7-4e14-9ee5-b3d8971c6de0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.598087] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7485d9f-d553-47fe-89e5-d477f7393080 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.611794] env[62600]: DEBUG nova.compute.provider_tree [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1170.663620] env[62600]: DEBUG oslo_vmware.api [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Task: {'id': task-1223096, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145644} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.663894] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1170.664105] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1170.664310] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1170.664487] env[62600]: INFO nova.compute.manager [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1170.664725] env[62600]: DEBUG oslo.service.loopingcall [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1170.664949] env[62600]: DEBUG nova.compute.manager [-] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1170.665048] env[62600]: DEBUG nova.network.neutron [-] [instance: 80a9de40-675d-456b-a91b-432d2315144b] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1171.115045] env[62600]: DEBUG nova.scheduler.client.report [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1171.126035] env[62600]: DEBUG nova.compute.manager [req-7b86be5a-13a2-49e7-bc88-cb702740b38b req-38d0ed0f-7a55-4a11-9185-eeab054a0a10 service nova] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Received event network-vif-deleted-4cf97208-acde-4474-94d5-27c1ed6ddd8c {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1171.126240] env[62600]: INFO nova.compute.manager [req-7b86be5a-13a2-49e7-bc88-cb702740b38b req-38d0ed0f-7a55-4a11-9185-eeab054a0a10 service nova] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Neutron deleted interface 4cf97208-acde-4474-94d5-27c1ed6ddd8c; detaching it from the instance and deleting it from the info cache [ 1171.126410] env[62600]: DEBUG nova.network.neutron [req-7b86be5a-13a2-49e7-bc88-cb702740b38b req-38d0ed0f-7a55-4a11-9185-eeab054a0a10 service nova] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1171.605822] env[62600]: DEBUG nova.network.neutron [-] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1171.620667] env[62600]: DEBUG oslo_concurrency.lockutils [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.616s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.629039] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5304381a-14f1-4ba1-bcb8-3eb2ef720905 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.640097] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b2bbf51-9906-4fea-8956-bee39e569bc0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.652230] env[62600]: INFO nova.scheduler.client.report [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Deleted allocations for instance c2f38d56-72fa-4bf0-a2cb-13d666571466 [ 1171.669630] env[62600]: DEBUG nova.compute.manager [req-7b86be5a-13a2-49e7-bc88-cb702740b38b req-38d0ed0f-7a55-4a11-9185-eeab054a0a10 service nova] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Detach interface failed, port_id=4cf97208-acde-4474-94d5-27c1ed6ddd8c, reason: Instance 80a9de40-675d-456b-a91b-432d2315144b could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1172.109121] env[62600]: INFO nova.compute.manager [-] [instance: 80a9de40-675d-456b-a91b-432d2315144b] Took 1.44 seconds to deallocate network for instance. [ 1172.159923] env[62600]: DEBUG oslo_concurrency.lockutils [None req-097b38e1-507d-46c8-b6ad-323661e3f591 tempest-ServersV294TestFqdnHostnames-1773242219 tempest-ServersV294TestFqdnHostnames-1773242219-project-member] Lock "c2f38d56-72fa-4bf0-a2cb-13d666571466" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.197s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.615167] env[62600]: DEBUG oslo_concurrency.lockutils [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.615532] env[62600]: DEBUG oslo_concurrency.lockutils [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.615667] env[62600]: DEBUG nova.objects.instance [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lazy-loading 'resources' on Instance uuid 80a9de40-675d-456b-a91b-432d2315144b {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1173.169018] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d449b1db-8d22-4b43-87e3-29990ec6dc86 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.178191] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-102df34b-47ca-4ddf-9668-eb379508ccfb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.208136] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b892fc7-20ef-4c86-bd2b-d2e498060a2e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.215323] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae3aad35-0ff0-4177-87ba-29ffb05b739e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.228343] env[62600]: DEBUG nova.compute.provider_tree [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1173.732037] env[62600]: DEBUG nova.scheduler.client.report [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1174.236483] env[62600]: DEBUG oslo_concurrency.lockutils [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.621s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.259810] env[62600]: INFO nova.scheduler.client.report [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Deleted allocations for instance 80a9de40-675d-456b-a91b-432d2315144b [ 1174.563047] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4ffb23d7-4c01-47f7-bcd6-ed4f13cf4e1b tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "692fa608-39f4-4d7d-9f30-b6559878b970" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1174.563389] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4ffb23d7-4c01-47f7-bcd6-ed4f13cf4e1b tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "692fa608-39f4-4d7d-9f30-b6559878b970" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1174.767672] env[62600]: DEBUG oslo_concurrency.lockutils [None req-bebb10f1-65a2-4229-9129-e7afe4b61475 tempest-AttachVolumeNegativeTest-428987982 tempest-AttachVolumeNegativeTest-428987982-project-member] Lock "80a9de40-675d-456b-a91b-432d2315144b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.216s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.066808] env[62600]: INFO nova.compute.manager [None req-4ffb23d7-4c01-47f7-bcd6-ed4f13cf4e1b tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Detaching volume 34095f5d-35c4-4bad-88a3-e3e627c33b5c [ 1175.102040] env[62600]: INFO nova.virt.block_device [None req-4ffb23d7-4c01-47f7-bcd6-ed4f13cf4e1b tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Attempting to driver detach volume 34095f5d-35c4-4bad-88a3-e3e627c33b5c from mountpoint /dev/sdb [ 1175.102040] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ffb23d7-4c01-47f7-bcd6-ed4f13cf4e1b tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Volume detach. Driver type: vmdk {{(pid=62600) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1175.102040] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ffb23d7-4c01-47f7-bcd6-ed4f13cf4e1b tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264377', 'volume_id': '34095f5d-35c4-4bad-88a3-e3e627c33b5c', 'name': 'volume-34095f5d-35c4-4bad-88a3-e3e627c33b5c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '692fa608-39f4-4d7d-9f30-b6559878b970', 'attached_at': '', 'detached_at': '', 'volume_id': '34095f5d-35c4-4bad-88a3-e3e627c33b5c', 'serial': '34095f5d-35c4-4bad-88a3-e3e627c33b5c'} {{(pid=62600) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1175.102040] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9876ad20-b262-4667-829e-4c4c008bdf96 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.135692] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e4036f-d116-43be-9ac5-3900e1d1976f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.145310] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-874b93ed-5234-484d-b0e6-76456b2280eb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.167816] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca8d17a-e513-4e83-a379-5c4e7beb3329 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.183801] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ffb23d7-4c01-47f7-bcd6-ed4f13cf4e1b tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] The volume has not been displaced from its original location: [datastore1] volume-34095f5d-35c4-4bad-88a3-e3e627c33b5c/volume-34095f5d-35c4-4bad-88a3-e3e627c33b5c.vmdk. No consolidation needed. {{(pid=62600) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1175.189093] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ffb23d7-4c01-47f7-bcd6-ed4f13cf4e1b tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Reconfiguring VM instance instance-00000064 to detach disk 2001 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1175.189364] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a437c59-5fb7-49b0-b637-2accf93a6927 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.207225] env[62600]: DEBUG oslo_vmware.api [None req-4ffb23d7-4c01-47f7-bcd6-ed4f13cf4e1b tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1175.207225] env[62600]: value = "task-1223098" [ 1175.207225] env[62600]: _type = "Task" [ 1175.207225] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.218721] env[62600]: DEBUG oslo_vmware.api [None req-4ffb23d7-4c01-47f7-bcd6-ed4f13cf4e1b tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223098, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.716782] env[62600]: DEBUG oslo_vmware.api [None req-4ffb23d7-4c01-47f7-bcd6-ed4f13cf4e1b tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223098, 'name': ReconfigVM_Task, 'duration_secs': 0.220415} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.717019] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ffb23d7-4c01-47f7-bcd6-ed4f13cf4e1b tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Reconfigured VM instance instance-00000064 to detach disk 2001 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1175.721650] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-44f2d4b0-7bfa-40c8-bb84-9482c206886c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.736379] env[62600]: DEBUG oslo_vmware.api [None req-4ffb23d7-4c01-47f7-bcd6-ed4f13cf4e1b tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1175.736379] env[62600]: value = "task-1223099" [ 1175.736379] env[62600]: _type = "Task" [ 1175.736379] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.743759] env[62600]: DEBUG oslo_vmware.api [None req-4ffb23d7-4c01-47f7-bcd6-ed4f13cf4e1b tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223099, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.247597] env[62600]: DEBUG oslo_vmware.api [None req-4ffb23d7-4c01-47f7-bcd6-ed4f13cf4e1b tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223099, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.746857] env[62600]: DEBUG oslo_vmware.api [None req-4ffb23d7-4c01-47f7-bcd6-ed4f13cf4e1b tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223099, 'name': ReconfigVM_Task, 'duration_secs': 0.90694} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.747182] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ffb23d7-4c01-47f7-bcd6-ed4f13cf4e1b tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-264377', 'volume_id': '34095f5d-35c4-4bad-88a3-e3e627c33b5c', 'name': 'volume-34095f5d-35c4-4bad-88a3-e3e627c33b5c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '692fa608-39f4-4d7d-9f30-b6559878b970', 'attached_at': '', 'detached_at': '', 'volume_id': '34095f5d-35c4-4bad-88a3-e3e627c33b5c', 'serial': '34095f5d-35c4-4bad-88a3-e3e627c33b5c'} {{(pid=62600) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1176.991584] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "e1a6e69d-09d8-44f0-b3a6-405ab3820c07" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.991897] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lock "e1a6e69d-09d8-44f0-b3a6-405ab3820c07" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.292646] env[62600]: DEBUG nova.objects.instance [None req-4ffb23d7-4c01-47f7-bcd6-ed4f13cf4e1b tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lazy-loading 'flavor' on Instance uuid 692fa608-39f4-4d7d-9f30-b6559878b970 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1177.496353] env[62600]: DEBUG nova.compute.manager [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1178.019099] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1178.019385] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1178.020868] env[62600]: INFO nova.compute.claims [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1178.300668] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4ffb23d7-4c01-47f7-bcd6-ed4f13cf4e1b tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "692fa608-39f4-4d7d-9f30-b6559878b970" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.737s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.840703] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "692fa608-39f4-4d7d-9f30-b6559878b970" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1178.840976] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "692fa608-39f4-4d7d-9f30-b6559878b970" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1178.841224] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "692fa608-39f4-4d7d-9f30-b6559878b970-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1178.841421] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "692fa608-39f4-4d7d-9f30-b6559878b970-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1178.841603] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "692fa608-39f4-4d7d-9f30-b6559878b970-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.843724] env[62600]: INFO nova.compute.manager [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Terminating instance [ 1178.845522] env[62600]: DEBUG nova.compute.manager [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1178.845721] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1178.846570] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67dd4d4b-719c-4f24-be44-b953fc639119 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.854233] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1178.854454] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-45df09f0-92ad-4152-82f4-572affe2939c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.860544] env[62600]: DEBUG oslo_vmware.api [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1178.860544] env[62600]: value = "task-1223101" [ 1178.860544] env[62600]: _type = "Task" [ 1178.860544] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.867577] env[62600]: DEBUG oslo_vmware.api [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223101, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.064540] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103e9369-bb60-459f-82ed-b8769330ac4f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.071633] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb0ad10a-657b-48ba-b2d0-20ed86428cf1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.103288] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea44d33f-bdce-4053-a034-ce03bd4405d1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.110651] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06ecd3ca-433a-4b4a-b7b6-a263774d879c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.123921] env[62600]: DEBUG nova.compute.provider_tree [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1179.371154] env[62600]: DEBUG oslo_vmware.api [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223101, 'name': PowerOffVM_Task, 'duration_secs': 0.140519} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.371578] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1179.371578] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1179.371787] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d359f8a0-d251-4d2f-a354-b99fe764d247 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.427620] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1179.427858] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1179.428072] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Deleting the datastore file [datastore1] 692fa608-39f4-4d7d-9f30-b6559878b970 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1179.428342] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-60f7eef5-6e89-4b3c-905e-eef4a32fd83d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.434090] env[62600]: DEBUG oslo_vmware.api [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1179.434090] env[62600]: value = "task-1223104" [ 1179.434090] env[62600]: _type = "Task" [ 1179.434090] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.441526] env[62600]: DEBUG oslo_vmware.api [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223104, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.628542] env[62600]: DEBUG nova.scheduler.client.report [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1179.944449] env[62600]: DEBUG oslo_vmware.api [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223104, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.11428} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.944671] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1179.945024] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1179.945233] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1179.945420] env[62600]: INFO nova.compute.manager [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1179.945704] env[62600]: DEBUG oslo.service.loopingcall [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1179.945942] env[62600]: DEBUG nova.compute.manager [-] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1179.946059] env[62600]: DEBUG nova.network.neutron [-] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1180.133720] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.114s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1180.134364] env[62600]: DEBUG nova.compute.manager [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1180.545994] env[62600]: DEBUG nova.compute.manager [req-aab43836-0872-4b09-ae47-185c8aef01a5 req-1726c682-8b92-46f2-8b0a-1f526936ba36 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Received event network-vif-deleted-280baabd-3083-4f33-b294-1f988193665b {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1180.546287] env[62600]: INFO nova.compute.manager [req-aab43836-0872-4b09-ae47-185c8aef01a5 req-1726c682-8b92-46f2-8b0a-1f526936ba36 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Neutron deleted interface 280baabd-3083-4f33-b294-1f988193665b; detaching it from the instance and deleting it from the info cache [ 1180.546520] env[62600]: DEBUG nova.network.neutron [req-aab43836-0872-4b09-ae47-185c8aef01a5 req-1726c682-8b92-46f2-8b0a-1f526936ba36 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1180.639758] env[62600]: DEBUG nova.compute.utils [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1180.640775] env[62600]: DEBUG nova.compute.manager [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1180.640954] env[62600]: DEBUG nova.network.neutron [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1180.684070] env[62600]: DEBUG nova.policy [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03149ddca4f440b2b19d65fa53e7a8b2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3091650617964d398efef937402ebfdb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 1180.955506] env[62600]: DEBUG nova.network.neutron [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Successfully created port: 4de0ebd7-1530-47fd-8653-72b841a6d4d1 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1181.021496] env[62600]: DEBUG nova.network.neutron [-] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1181.050018] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0f9256ca-32ca-4819-bbf9-095f4d473469 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.059011] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdce6b40-259d-426d-94d1-257863f5e0cb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.081008] env[62600]: DEBUG nova.compute.manager [req-aab43836-0872-4b09-ae47-185c8aef01a5 req-1726c682-8b92-46f2-8b0a-1f526936ba36 service nova] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Detach interface failed, port_id=280baabd-3083-4f33-b294-1f988193665b, reason: Instance 692fa608-39f4-4d7d-9f30-b6559878b970 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1181.146523] env[62600]: DEBUG nova.compute.manager [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1181.524790] env[62600]: INFO nova.compute.manager [-] [instance: 692fa608-39f4-4d7d-9f30-b6559878b970] Took 1.58 seconds to deallocate network for instance. [ 1182.031150] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1182.031488] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.031639] env[62600]: DEBUG nova.objects.instance [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lazy-loading 'resources' on Instance uuid 692fa608-39f4-4d7d-9f30-b6559878b970 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1182.155539] env[62600]: DEBUG nova.compute.manager [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1182.180406] env[62600]: DEBUG nova.virt.hardware [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1182.180654] env[62600]: DEBUG nova.virt.hardware [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1182.180816] env[62600]: DEBUG nova.virt.hardware [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1182.181009] env[62600]: DEBUG nova.virt.hardware [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1182.181174] env[62600]: DEBUG nova.virt.hardware [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1182.181327] env[62600]: DEBUG nova.virt.hardware [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1182.181543] env[62600]: DEBUG nova.virt.hardware [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1182.181707] env[62600]: DEBUG nova.virt.hardware [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1182.181896] env[62600]: DEBUG nova.virt.hardware [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1182.182091] env[62600]: DEBUG nova.virt.hardware [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1182.182308] env[62600]: DEBUG nova.virt.hardware [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1182.183162] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63eed86a-21c5-4610-9d98-9d4dda89cdf7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.191162] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e93911-a4d3-4710-b5a8-ad90d674fdf5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.453976] env[62600]: DEBUG nova.network.neutron [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Successfully updated port: 4de0ebd7-1530-47fd-8653-72b841a6d4d1 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1182.575659] env[62600]: DEBUG nova.compute.manager [req-a94adeb2-d689-4616-855d-6421be9dd210 req-89bfb599-0dd4-4c0c-8373-fd2d0b23a5d9 service nova] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Received event network-vif-plugged-4de0ebd7-1530-47fd-8653-72b841a6d4d1 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1182.575895] env[62600]: DEBUG oslo_concurrency.lockutils [req-a94adeb2-d689-4616-855d-6421be9dd210 req-89bfb599-0dd4-4c0c-8373-fd2d0b23a5d9 service nova] Acquiring lock "e1a6e69d-09d8-44f0-b3a6-405ab3820c07-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1182.576150] env[62600]: DEBUG oslo_concurrency.lockutils [req-a94adeb2-d689-4616-855d-6421be9dd210 req-89bfb599-0dd4-4c0c-8373-fd2d0b23a5d9 service nova] Lock "e1a6e69d-09d8-44f0-b3a6-405ab3820c07-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.576332] env[62600]: DEBUG oslo_concurrency.lockutils [req-a94adeb2-d689-4616-855d-6421be9dd210 req-89bfb599-0dd4-4c0c-8373-fd2d0b23a5d9 service nova] Lock "e1a6e69d-09d8-44f0-b3a6-405ab3820c07-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1182.576509] env[62600]: DEBUG nova.compute.manager [req-a94adeb2-d689-4616-855d-6421be9dd210 req-89bfb599-0dd4-4c0c-8373-fd2d0b23a5d9 service nova] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] No waiting events found dispatching network-vif-plugged-4de0ebd7-1530-47fd-8653-72b841a6d4d1 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1182.576855] env[62600]: WARNING nova.compute.manager [req-a94adeb2-d689-4616-855d-6421be9dd210 req-89bfb599-0dd4-4c0c-8373-fd2d0b23a5d9 service nova] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Received unexpected event network-vif-plugged-4de0ebd7-1530-47fd-8653-72b841a6d4d1 for instance with vm_state building and task_state spawning. [ 1182.577060] env[62600]: DEBUG nova.compute.manager [req-a94adeb2-d689-4616-855d-6421be9dd210 req-89bfb599-0dd4-4c0c-8373-fd2d0b23a5d9 service nova] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Received event network-changed-4de0ebd7-1530-47fd-8653-72b841a6d4d1 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1182.577282] env[62600]: DEBUG nova.compute.manager [req-a94adeb2-d689-4616-855d-6421be9dd210 req-89bfb599-0dd4-4c0c-8373-fd2d0b23a5d9 service nova] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Refreshing instance network info cache due to event network-changed-4de0ebd7-1530-47fd-8653-72b841a6d4d1. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1182.577434] env[62600]: DEBUG oslo_concurrency.lockutils [req-a94adeb2-d689-4616-855d-6421be9dd210 req-89bfb599-0dd4-4c0c-8373-fd2d0b23a5d9 service nova] Acquiring lock "refresh_cache-e1a6e69d-09d8-44f0-b3a6-405ab3820c07" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1182.577576] env[62600]: DEBUG oslo_concurrency.lockutils [req-a94adeb2-d689-4616-855d-6421be9dd210 req-89bfb599-0dd4-4c0c-8373-fd2d0b23a5d9 service nova] Acquired lock "refresh_cache-e1a6e69d-09d8-44f0-b3a6-405ab3820c07" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.577737] env[62600]: DEBUG nova.network.neutron [req-a94adeb2-d689-4616-855d-6421be9dd210 req-89bfb599-0dd4-4c0c-8373-fd2d0b23a5d9 service nova] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Refreshing network info cache for port 4de0ebd7-1530-47fd-8653-72b841a6d4d1 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1182.580246] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea179e9c-ae8f-4c0d-9458-b1fce1e99754 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.589876] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088c0790-c68c-4b46-820d-1e9c2dbd4388 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.620048] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0ab74aa-a1f4-44f8-be85-20cf09f04b2f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.626743] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c50cbff7-cbc9-430d-b951-2346fb114be4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.640151] env[62600]: DEBUG nova.compute.provider_tree [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1182.956670] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "refresh_cache-e1a6e69d-09d8-44f0-b3a6-405ab3820c07" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1183.113979] env[62600]: DEBUG nova.network.neutron [req-a94adeb2-d689-4616-855d-6421be9dd210 req-89bfb599-0dd4-4c0c-8373-fd2d0b23a5d9 service nova] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1183.143534] env[62600]: DEBUG nova.scheduler.client.report [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1183.186633] env[62600]: DEBUG nova.network.neutron [req-a94adeb2-d689-4616-855d-6421be9dd210 req-89bfb599-0dd4-4c0c-8373-fd2d0b23a5d9 service nova] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1183.650041] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.617s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.671302] env[62600]: INFO nova.scheduler.client.report [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Deleted allocations for instance 692fa608-39f4-4d7d-9f30-b6559878b970 [ 1183.689803] env[62600]: DEBUG oslo_concurrency.lockutils [req-a94adeb2-d689-4616-855d-6421be9dd210 req-89bfb599-0dd4-4c0c-8373-fd2d0b23a5d9 service nova] Releasing lock "refresh_cache-e1a6e69d-09d8-44f0-b3a6-405ab3820c07" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1183.690251] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquired lock "refresh_cache-e1a6e69d-09d8-44f0-b3a6-405ab3820c07" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.690425] env[62600]: DEBUG nova.network.neutron [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1184.180817] env[62600]: DEBUG oslo_concurrency.lockutils [None req-1207c42d-b68c-421b-9954-907d78a47c51 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "692fa608-39f4-4d7d-9f30-b6559878b970" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.340s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.220618] env[62600]: DEBUG nova.network.neutron [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1184.331120] env[62600]: DEBUG nova.network.neutron [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Updating instance_info_cache with network_info: [{"id": "4de0ebd7-1530-47fd-8653-72b841a6d4d1", "address": "fa:16:3e:7e:4a:c8", "network": {"id": "bcd55d3c-9d84-4e1d-b1fd-d6f1ac0cff02", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-771276928-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3091650617964d398efef937402ebfdb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4de0ebd7-15", "ovs_interfaceid": "4de0ebd7-1530-47fd-8653-72b841a6d4d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1184.833657] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Releasing lock "refresh_cache-e1a6e69d-09d8-44f0-b3a6-405ab3820c07" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1184.834026] env[62600]: DEBUG nova.compute.manager [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Instance network_info: |[{"id": "4de0ebd7-1530-47fd-8653-72b841a6d4d1", "address": "fa:16:3e:7e:4a:c8", "network": {"id": "bcd55d3c-9d84-4e1d-b1fd-d6f1ac0cff02", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-771276928-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3091650617964d398efef937402ebfdb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4de0ebd7-15", "ovs_interfaceid": "4de0ebd7-1530-47fd-8653-72b841a6d4d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1184.835111] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7e:4a:c8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a10c88d7-d13f-44fd-acee-7a734eb5f56a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4de0ebd7-1530-47fd-8653-72b841a6d4d1', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1184.842285] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Creating folder: Project (3091650617964d398efef937402ebfdb). Parent ref: group-v264198. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1184.842580] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1e60db8c-cd26-4387-bb76-e74d00f0a372 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.853300] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Created folder: Project (3091650617964d398efef937402ebfdb) in parent group-v264198. [ 1184.853484] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Creating folder: Instances. Parent ref: group-v264381. {{(pid=62600) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1184.853767] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-94244a8a-9bda-4999-a7d4-3ca19e3eb3e3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.862541] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Created folder: Instances in parent group-v264381. [ 1184.862782] env[62600]: DEBUG oslo.service.loopingcall [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1184.862977] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1184.863190] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-904c751c-96b4-4988-bd81-d6c20f65199f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.881263] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1184.881263] env[62600]: value = "task-1223107" [ 1184.881263] env[62600]: _type = "Task" [ 1184.881263] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.888055] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1223107, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.391408] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1223107, 'name': CreateVM_Task, 'duration_secs': 0.28173} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.391726] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1185.392298] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1185.392478] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.392807] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1185.393066] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14c33093-1c6f-4eda-8810-e48b5dc05a37 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.397146] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1185.397146] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d08300-db0c-91ed-54e5-47b5b056b7dd" [ 1185.397146] env[62600]: _type = "Task" [ 1185.397146] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.404585] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d08300-db0c-91ed-54e5-47b5b056b7dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.750578] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "00648d57-68ae-438b-af37-7e50d6cdcedd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.750828] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "00648d57-68ae-438b-af37-7e50d6cdcedd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.907426] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52d08300-db0c-91ed-54e5-47b5b056b7dd, 'name': SearchDatastore_Task, 'duration_secs': 0.008825} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.907725] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1185.907957] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1185.908207] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1185.908357] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.908538] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1185.908779] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d238dbfc-096f-4e22-8acb-6c716059c33e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.916323] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1185.916499] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1185.917178] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e32184a0-2856-4ea7-bfeb-18700d97ad83 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.921662] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1185.921662] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520cb565-3b55-7c62-85c7-c1d650d32f84" [ 1185.921662] env[62600]: _type = "Task" [ 1185.921662] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.928297] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520cb565-3b55-7c62-85c7-c1d650d32f84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.253612] env[62600]: DEBUG nova.compute.manager [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1186.432417] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520cb565-3b55-7c62-85c7-c1d650d32f84, 'name': SearchDatastore_Task, 'duration_secs': 0.007191} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.433243] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6891e1e-63e8-493b-a0a3-ce799eb2c9e0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.438397] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1186.438397] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]527bd92b-fbf8-e689-7fd4-0ad330080ac0" [ 1186.438397] env[62600]: _type = "Task" [ 1186.438397] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.445288] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]527bd92b-fbf8-e689-7fd4-0ad330080ac0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.772487] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1186.772753] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1186.774280] env[62600]: INFO nova.compute.claims [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1186.952300] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]527bd92b-fbf8-e689-7fd4-0ad330080ac0, 'name': SearchDatastore_Task, 'duration_secs': 0.008595} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.952574] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1186.952853] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] e1a6e69d-09d8-44f0-b3a6-405ab3820c07/e1a6e69d-09d8-44f0-b3a6-405ab3820c07.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1186.953155] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fd5207dc-5471-4525-ada1-f56446433e8f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.959270] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1186.959270] env[62600]: value = "task-1223108" [ 1186.959270] env[62600]: _type = "Task" [ 1186.959270] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.966295] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223108, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.468837] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223108, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.424282} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.469196] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] e1a6e69d-09d8-44f0-b3a6-405ab3820c07/e1a6e69d-09d8-44f0-b3a6-405ab3820c07.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1187.469359] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1187.469622] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-42547696-321f-4b38-9c37-a5f63a343284 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.477074] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1187.477074] env[62600]: value = "task-1223109" [ 1187.477074] env[62600]: _type = "Task" [ 1187.477074] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.484450] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223109, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.817681] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8034cfb0-7b33-41fb-85a9-0c3220ffceaf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.825155] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b9d2eed-0f33-4074-a3ca-d99f51130ef5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.853543] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38724a25-01b8-47d3-bda3-ceec17d81f85 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.860186] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-708ffab3-fd77-4214-be2e-e52904b8e9d7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.872584] env[62600]: DEBUG nova.compute.provider_tree [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1187.987794] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223109, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057265} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.988066] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1187.988792] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07ec6f66-bf22-4bbb-9c46-eddd5da0588b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.009634] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] e1a6e69d-09d8-44f0-b3a6-405ab3820c07/e1a6e69d-09d8-44f0-b3a6-405ab3820c07.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1188.010141] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ffe081e-1f95-4450-a637-6b80995b1689 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.029289] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1188.029289] env[62600]: value = "task-1223110" [ 1188.029289] env[62600]: _type = "Task" [ 1188.029289] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.036725] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223110, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.376130] env[62600]: DEBUG nova.scheduler.client.report [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1188.538663] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223110, 'name': ReconfigVM_Task, 'duration_secs': 0.276761} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.538963] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Reconfigured VM instance instance-00000069 to attach disk [datastore2] e1a6e69d-09d8-44f0-b3a6-405ab3820c07/e1a6e69d-09d8-44f0-b3a6-405ab3820c07.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1188.539590] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5b8984bf-8767-427c-a474-01886df3d719 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.545443] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1188.545443] env[62600]: value = "task-1223111" [ 1188.545443] env[62600]: _type = "Task" [ 1188.545443] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.552424] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223111, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.881470] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.108s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1188.881955] env[62600]: DEBUG nova.compute.manager [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1189.055648] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223111, 'name': Rename_Task, 'duration_secs': 0.134222} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.055940] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1189.056217] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d9dceda2-6356-4df2-bcf3-130ac722e0d8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.062471] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1189.062471] env[62600]: value = "task-1223112" [ 1189.062471] env[62600]: _type = "Task" [ 1189.062471] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.069296] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223112, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.387108] env[62600]: DEBUG nova.compute.utils [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1189.389057] env[62600]: DEBUG nova.compute.manager [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1189.389057] env[62600]: DEBUG nova.network.neutron [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1189.426981] env[62600]: DEBUG nova.policy [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '57644cb1856e4efba88d25065ce1dc68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '952323d62f034126a9b46750d03fe6ea', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 1189.572926] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223112, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.697541] env[62600]: DEBUG nova.network.neutron [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Successfully created port: 9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1189.891620] env[62600]: DEBUG nova.compute.manager [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1190.072869] env[62600]: DEBUG oslo_vmware.api [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223112, 'name': PowerOnVM_Task, 'duration_secs': 0.518907} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.073222] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1190.073431] env[62600]: INFO nova.compute.manager [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Took 7.92 seconds to spawn the instance on the hypervisor. [ 1190.073616] env[62600]: DEBUG nova.compute.manager [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1190.074378] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cd4bfc3-103d-4be4-a6fd-120866fdbf3b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.590555] env[62600]: INFO nova.compute.manager [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Took 12.59 seconds to build instance. [ 1190.901232] env[62600]: DEBUG nova.compute.manager [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1190.925951] env[62600]: DEBUG nova.virt.hardware [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1190.926253] env[62600]: DEBUG nova.virt.hardware [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1190.926404] env[62600]: DEBUG nova.virt.hardware [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1190.926663] env[62600]: DEBUG nova.virt.hardware [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1190.926826] env[62600]: DEBUG nova.virt.hardware [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1190.926980] env[62600]: DEBUG nova.virt.hardware [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1190.927217] env[62600]: DEBUG nova.virt.hardware [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1190.927425] env[62600]: DEBUG nova.virt.hardware [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1190.927605] env[62600]: DEBUG nova.virt.hardware [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1190.927778] env[62600]: DEBUG nova.virt.hardware [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1190.927966] env[62600]: DEBUG nova.virt.hardware [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1190.928831] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76a6e0fe-50d7-4d6a-ae6a-c35e0410154a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.936720] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b669f8a-d074-4631-837b-6a921a20ac60 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.070985] env[62600]: DEBUG nova.compute.manager [req-ddd9206d-e8cc-41d4-89b5-d6b75008a4c9 req-fba309c7-8442-4e81-b31f-3f61a4943009 service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Received event network-vif-plugged-9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1191.070985] env[62600]: DEBUG oslo_concurrency.lockutils [req-ddd9206d-e8cc-41d4-89b5-d6b75008a4c9 req-fba309c7-8442-4e81-b31f-3f61a4943009 service nova] Acquiring lock "00648d57-68ae-438b-af37-7e50d6cdcedd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.070985] env[62600]: DEBUG oslo_concurrency.lockutils [req-ddd9206d-e8cc-41d4-89b5-d6b75008a4c9 req-fba309c7-8442-4e81-b31f-3f61a4943009 service nova] Lock "00648d57-68ae-438b-af37-7e50d6cdcedd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.071354] env[62600]: DEBUG oslo_concurrency.lockutils [req-ddd9206d-e8cc-41d4-89b5-d6b75008a4c9 req-fba309c7-8442-4e81-b31f-3f61a4943009 service nova] Lock "00648d57-68ae-438b-af37-7e50d6cdcedd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.071415] env[62600]: DEBUG nova.compute.manager [req-ddd9206d-e8cc-41d4-89b5-d6b75008a4c9 req-fba309c7-8442-4e81-b31f-3f61a4943009 service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] No waiting events found dispatching network-vif-plugged-9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1191.072702] env[62600]: WARNING nova.compute.manager [req-ddd9206d-e8cc-41d4-89b5-d6b75008a4c9 req-fba309c7-8442-4e81-b31f-3f61a4943009 service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Received unexpected event network-vif-plugged-9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 for instance with vm_state building and task_state spawning. [ 1191.093039] env[62600]: DEBUG oslo_concurrency.lockutils [None req-85c8e1de-db14-4d5e-b1f9-14407950d69b tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lock "e1a6e69d-09d8-44f0-b3a6-405ab3820c07" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.100s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.160136] env[62600]: DEBUG nova.network.neutron [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Successfully updated port: 9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1191.577775] env[62600]: INFO nova.compute.manager [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Rescuing [ 1191.578023] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "refresh_cache-e1a6e69d-09d8-44f0-b3a6-405ab3820c07" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1191.578226] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquired lock "refresh_cache-e1a6e69d-09d8-44f0-b3a6-405ab3820c07" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1191.578377] env[62600]: DEBUG nova.network.neutron [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1191.662655] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1191.662968] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquired lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1191.662968] env[62600]: DEBUG nova.network.neutron [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1192.195164] env[62600]: DEBUG nova.network.neutron [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1192.276040] env[62600]: DEBUG nova.network.neutron [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Updating instance_info_cache with network_info: [{"id": "4de0ebd7-1530-47fd-8653-72b841a6d4d1", "address": "fa:16:3e:7e:4a:c8", "network": {"id": "bcd55d3c-9d84-4e1d-b1fd-d6f1ac0cff02", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-771276928-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3091650617964d398efef937402ebfdb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4de0ebd7-15", "ovs_interfaceid": "4de0ebd7-1530-47fd-8653-72b841a6d4d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1192.371040] env[62600]: DEBUG nova.network.neutron [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Updating instance_info_cache with network_info: [{"id": "9c4c9eaa-2ecd-4598-bed6-f746cbe53f80", "address": "fa:16:3e:6b:09:8c", "network": {"id": "dc21424f-b891-428d-a74e-76ab78d757a8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1428037886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "952323d62f034126a9b46750d03fe6ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c4c9eaa-2e", "ovs_interfaceid": "9c4c9eaa-2ecd-4598-bed6-f746cbe53f80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1192.779544] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Releasing lock "refresh_cache-e1a6e69d-09d8-44f0-b3a6-405ab3820c07" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1192.873224] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Releasing lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1192.873550] env[62600]: DEBUG nova.compute.manager [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Instance network_info: |[{"id": "9c4c9eaa-2ecd-4598-bed6-f746cbe53f80", "address": "fa:16:3e:6b:09:8c", "network": {"id": "dc21424f-b891-428d-a74e-76ab78d757a8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1428037886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "952323d62f034126a9b46750d03fe6ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c4c9eaa-2e", "ovs_interfaceid": "9c4c9eaa-2ecd-4598-bed6-f746cbe53f80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1192.873985] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:09:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '07e9bef1-2b0e-4e4d-997f-de71bb0e213a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9c4c9eaa-2ecd-4598-bed6-f746cbe53f80', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1192.881607] env[62600]: DEBUG oslo.service.loopingcall [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1192.881818] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1192.882061] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b4099c1e-4a88-482e-91b5-f9112d37ba42 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.903194] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1192.903194] env[62600]: value = "task-1223113" [ 1192.903194] env[62600]: _type = "Task" [ 1192.903194] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.911852] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1223113, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.096788] env[62600]: DEBUG nova.compute.manager [req-d31c0509-def6-44e0-85a1-c56f0ba25803 req-cba688e6-d3f5-4b88-b920-12635a71f8c9 service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Received event network-changed-9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1193.096970] env[62600]: DEBUG nova.compute.manager [req-d31c0509-def6-44e0-85a1-c56f0ba25803 req-cba688e6-d3f5-4b88-b920-12635a71f8c9 service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Refreshing instance network info cache due to event network-changed-9c4c9eaa-2ecd-4598-bed6-f746cbe53f80. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1193.097164] env[62600]: DEBUG oslo_concurrency.lockutils [req-d31c0509-def6-44e0-85a1-c56f0ba25803 req-cba688e6-d3f5-4b88-b920-12635a71f8c9 service nova] Acquiring lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1193.097319] env[62600]: DEBUG oslo_concurrency.lockutils [req-d31c0509-def6-44e0-85a1-c56f0ba25803 req-cba688e6-d3f5-4b88-b920-12635a71f8c9 service nova] Acquired lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1193.097485] env[62600]: DEBUG nova.network.neutron [req-d31c0509-def6-44e0-85a1-c56f0ba25803 req-cba688e6-d3f5-4b88-b920-12635a71f8c9 service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Refreshing network info cache for port 9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1193.305061] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1193.305285] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6fcb9a09-c339-4d91-848d-3174cded6a79 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.312119] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1193.312119] env[62600]: value = "task-1223114" [ 1193.312119] env[62600]: _type = "Task" [ 1193.312119] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.319807] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223114, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.414286] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1223113, 'name': CreateVM_Task, 'duration_secs': 0.351457} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.414573] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1193.415487] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1193.415765] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1193.416153] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1193.416448] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-000c7fa3-51e0-46ae-8bed-92ed3db7f1e2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.421474] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1193.421474] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5272756b-eda0-05a4-aed1-4af5977051ec" [ 1193.421474] env[62600]: _type = "Task" [ 1193.421474] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.429172] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5272756b-eda0-05a4-aed1-4af5977051ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.773950] env[62600]: DEBUG nova.network.neutron [req-d31c0509-def6-44e0-85a1-c56f0ba25803 req-cba688e6-d3f5-4b88-b920-12635a71f8c9 service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Updated VIF entry in instance network info cache for port 9c4c9eaa-2ecd-4598-bed6-f746cbe53f80. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1193.774335] env[62600]: DEBUG nova.network.neutron [req-d31c0509-def6-44e0-85a1-c56f0ba25803 req-cba688e6-d3f5-4b88-b920-12635a71f8c9 service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Updating instance_info_cache with network_info: [{"id": "9c4c9eaa-2ecd-4598-bed6-f746cbe53f80", "address": "fa:16:3e:6b:09:8c", "network": {"id": "dc21424f-b891-428d-a74e-76ab78d757a8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1428037886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "952323d62f034126a9b46750d03fe6ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c4c9eaa-2e", "ovs_interfaceid": "9c4c9eaa-2ecd-4598-bed6-f746cbe53f80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1193.822232] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223114, 'name': PowerOffVM_Task, 'duration_secs': 0.168457} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.822498] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1193.823255] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c4e4ba8-9eed-4c8e-97fe-c32d1cdc133e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.840523] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbd6a1d5-f3ba-41af-8182-0446f5314c49 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.865942] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1193.866249] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3bdcfd95-5263-479c-8ea4-14cbb23c320e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.872794] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1193.872794] env[62600]: value = "task-1223115" [ 1193.872794] env[62600]: _type = "Task" [ 1193.872794] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.879677] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223115, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.930763] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5272756b-eda0-05a4-aed1-4af5977051ec, 'name': SearchDatastore_Task, 'duration_secs': 0.013048} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.931064] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1193.931304] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1193.931539] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1193.931689] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1193.931869] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1193.932123] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-edac0184-7721-4898-9c66-19864ad6b1bd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.940081] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1193.940268] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1193.940985] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7d180f1-479d-4f44-8cc8-3525d91ede96 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.945464] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1193.945464] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f42502-c8c9-7cfc-4ce7-6c98d6b82b83" [ 1193.945464] env[62600]: _type = "Task" [ 1193.945464] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.952273] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f42502-c8c9-7cfc-4ce7-6c98d6b82b83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.277559] env[62600]: DEBUG oslo_concurrency.lockutils [req-d31c0509-def6-44e0-85a1-c56f0ba25803 req-cba688e6-d3f5-4b88-b920-12635a71f8c9 service nova] Releasing lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1194.384128] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] VM already powered off {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1194.384345] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1194.384562] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1194.454288] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52f42502-c8c9-7cfc-4ce7-6c98d6b82b83, 'name': SearchDatastore_Task, 'duration_secs': 0.007898} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.455043] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab2697f4-e95f-4cce-92dc-dc6a55e2214b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.459650] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1194.459650] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52bab9f1-b62b-4af0-c08a-aa7574135a2b" [ 1194.459650] env[62600]: _type = "Task" [ 1194.459650] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.466489] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52bab9f1-b62b-4af0-c08a-aa7574135a2b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.970075] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52bab9f1-b62b-4af0-c08a-aa7574135a2b, 'name': SearchDatastore_Task, 'duration_secs': 0.008456} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.970373] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1194.970618] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 00648d57-68ae-438b-af37-7e50d6cdcedd/00648d57-68ae-438b-af37-7e50d6cdcedd.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1194.970889] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1194.971090] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1194.971299] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1ce77338-021c-41e9-9eb7-d9dac6929714 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.973097] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2953bccd-7edf-44a3-85dd-3159b3f29dbf {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.981805] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1194.982016] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1194.983368] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03de5796-def5-424e-872c-8235413f69a3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.985417] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1194.985417] env[62600]: value = "task-1223116" [ 1194.985417] env[62600]: _type = "Task" [ 1194.985417] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.989224] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1194.989224] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528bb823-504e-34c3-cca4-b78a78d3b282" [ 1194.989224] env[62600]: _type = "Task" [ 1194.989224] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.994826] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223116, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.999540] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]528bb823-504e-34c3-cca4-b78a78d3b282, 'name': SearchDatastore_Task, 'duration_secs': 0.007558} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.000231] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69dea145-f94a-4423-8c72-52f0155698e2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.004608] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1195.004608] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e4337e-6db9-b597-c9f8-fcb9c084a5af" [ 1195.004608] env[62600]: _type = "Task" [ 1195.004608] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.011439] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e4337e-6db9-b597-c9f8-fcb9c084a5af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.495592] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223116, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.43176} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.495843] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] 00648d57-68ae-438b-af37-7e50d6cdcedd/00648d57-68ae-438b-af37-7e50d6cdcedd.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1195.496051] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1195.496304] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-106882ba-8715-4eb4-9f78-cd4e4ab7e20d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.502699] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1195.502699] env[62600]: value = "task-1223117" [ 1195.502699] env[62600]: _type = "Task" [ 1195.502699] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.512367] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223117, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.515698] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52e4337e-6db9-b597-c9f8-fcb9c084a5af, 'name': SearchDatastore_Task, 'duration_secs': 0.006966} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.515952] env[62600]: DEBUG oslo_concurrency.lockutils [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1195.516272] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] e1a6e69d-09d8-44f0-b3a6-405ab3820c07/e7bd1d77-bfff-4684-9545-f6a1c69efe58-rescue.vmdk. {{(pid=62600) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1195.516532] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d74f4655-3c9b-40de-8e2d-0220fdb876e9 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.521772] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1195.521772] env[62600]: value = "task-1223118" [ 1195.521772] env[62600]: _type = "Task" [ 1195.521772] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.529254] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223118, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.012489] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223117, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067959} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.012839] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1196.013460] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef3a54e-7f30-4452-9916-e864ff06747b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.034466] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 00648d57-68ae-438b-af37-7e50d6cdcedd/00648d57-68ae-438b-af37-7e50d6cdcedd.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1196.037230] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7599576e-c661-4eb0-8fd1-151c3a07f87d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.055768] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223118, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.385914} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.056871] env[62600]: INFO nova.virt.vmwareapi.ds_util [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore2] e1a6e69d-09d8-44f0-b3a6-405ab3820c07/e7bd1d77-bfff-4684-9545-f6a1c69efe58-rescue.vmdk. [ 1196.057210] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1196.057210] env[62600]: value = "task-1223119" [ 1196.057210] env[62600]: _type = "Task" [ 1196.057210] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.057860] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-852b80e4-a0fc-4d22-9fb7-70c3c56142dc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.068568] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223119, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.087719] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] e1a6e69d-09d8-44f0-b3a6-405ab3820c07/e7bd1d77-bfff-4684-9545-f6a1c69efe58-rescue.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1196.087982] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eca23726-9d42-4d41-88de-1800b01bfb5b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.105445] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1196.105445] env[62600]: value = "task-1223120" [ 1196.105445] env[62600]: _type = "Task" [ 1196.105445] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.113093] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223120, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.569544] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223119, 'name': ReconfigVM_Task, 'duration_secs': 0.302357} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.569836] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 00648d57-68ae-438b-af37-7e50d6cdcedd/00648d57-68ae-438b-af37-7e50d6cdcedd.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1196.570491] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bb440d30-d6d1-425f-9283-c0a0b0af7898 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.576599] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1196.576599] env[62600]: value = "task-1223121" [ 1196.576599] env[62600]: _type = "Task" [ 1196.576599] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.583681] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223121, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.613840] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223120, 'name': ReconfigVM_Task, 'duration_secs': 0.317729} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.614145] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Reconfigured VM instance instance-00000069 to attach disk [datastore2] e1a6e69d-09d8-44f0-b3a6-405ab3820c07/e7bd1d77-bfff-4684-9545-f6a1c69efe58-rescue.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1196.614988] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-551ca77a-16af-4d31-b44a-8c9d19701559 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.638769] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f299fa54-3c26-4129-b0f4-e24d6ab737c8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.653218] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1196.653218] env[62600]: value = "task-1223122" [ 1196.653218] env[62600]: _type = "Task" [ 1196.653218] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.660420] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223122, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.086745] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223121, 'name': Rename_Task, 'duration_secs': 0.132544} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.087103] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1197.087279] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5eb8be38-81b2-4b55-81cb-6be429da4bb8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.093378] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1197.093378] env[62600]: value = "task-1223123" [ 1197.093378] env[62600]: _type = "Task" [ 1197.093378] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.100293] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223123, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.161574] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223122, 'name': ReconfigVM_Task, 'duration_secs': 0.154142} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.161880] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1197.162182] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fecba9a4-c4fa-4164-9191-301552412cd7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.167916] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1197.167916] env[62600]: value = "task-1223124" [ 1197.167916] env[62600]: _type = "Task" [ 1197.167916] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.174779] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223124, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.603344] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223123, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.676846] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223124, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.104772] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223123, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.178592] env[62600]: DEBUG oslo_vmware.api [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223124, 'name': PowerOnVM_Task, 'duration_secs': 0.981243} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.178806] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1198.181627] env[62600]: DEBUG nova.compute.manager [None req-67905a50-cbc4-464d-8e45-b64484b8972f tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1198.182380] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6d085b8-d4fd-40f1-8e72-1bec7770db80 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.605159] env[62600]: DEBUG oslo_vmware.api [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223123, 'name': PowerOnVM_Task, 'duration_secs': 1.099648} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.605436] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1198.605647] env[62600]: INFO nova.compute.manager [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Took 7.70 seconds to spawn the instance on the hypervisor. [ 1198.605833] env[62600]: DEBUG nova.compute.manager [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1198.606613] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f5e84b4-0aaf-448a-b350-f1122f70dc25 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.123621] env[62600]: INFO nova.compute.manager [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Took 12.37 seconds to build instance. [ 1199.625843] env[62600]: DEBUG oslo_concurrency.lockutils [None req-12e034fb-a57d-465e-837b-5194fb4d78cf tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "00648d57-68ae-438b-af37-7e50d6cdcedd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.875s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1199.960283] env[62600]: DEBUG nova.compute.manager [req-ae5a90ca-b359-4e22-b8b6-462b0f7d1b58 req-5af67daf-033d-46f8-baf9-d6c49cf0ba8a service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Received event network-changed-9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1199.960620] env[62600]: DEBUG nova.compute.manager [req-ae5a90ca-b359-4e22-b8b6-462b0f7d1b58 req-5af67daf-033d-46f8-baf9-d6c49cf0ba8a service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Refreshing instance network info cache due to event network-changed-9c4c9eaa-2ecd-4598-bed6-f746cbe53f80. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1199.960706] env[62600]: DEBUG oslo_concurrency.lockutils [req-ae5a90ca-b359-4e22-b8b6-462b0f7d1b58 req-5af67daf-033d-46f8-baf9-d6c49cf0ba8a service nova] Acquiring lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.960892] env[62600]: DEBUG oslo_concurrency.lockutils [req-ae5a90ca-b359-4e22-b8b6-462b0f7d1b58 req-5af67daf-033d-46f8-baf9-d6c49cf0ba8a service nova] Acquired lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.961053] env[62600]: DEBUG nova.network.neutron [req-ae5a90ca-b359-4e22-b8b6-462b0f7d1b58 req-5af67daf-033d-46f8-baf9-d6c49cf0ba8a service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Refreshing network info cache for port 9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1200.010036] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "d34b306a-ba12-4923-b3e3-76db3bc1cd9e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1200.010295] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lock "d34b306a-ba12-4923-b3e3-76db3bc1cd9e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1200.513034] env[62600]: DEBUG nova.compute.manager [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Starting instance... {{(pid=62600) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1200.680371] env[62600]: DEBUG nova.network.neutron [req-ae5a90ca-b359-4e22-b8b6-462b0f7d1b58 req-5af67daf-033d-46f8-baf9-d6c49cf0ba8a service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Updated VIF entry in instance network info cache for port 9c4c9eaa-2ecd-4598-bed6-f746cbe53f80. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1200.680817] env[62600]: DEBUG nova.network.neutron [req-ae5a90ca-b359-4e22-b8b6-462b0f7d1b58 req-5af67daf-033d-46f8-baf9-d6c49cf0ba8a service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Updating instance_info_cache with network_info: [{"id": "9c4c9eaa-2ecd-4598-bed6-f746cbe53f80", "address": "fa:16:3e:6b:09:8c", "network": {"id": "dc21424f-b891-428d-a74e-76ab78d757a8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1428037886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "952323d62f034126a9b46750d03fe6ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c4c9eaa-2e", "ovs_interfaceid": "9c4c9eaa-2ecd-4598-bed6-f746cbe53f80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1201.037146] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1201.037431] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1201.038925] env[62600]: INFO nova.compute.claims [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1201.183312] env[62600]: DEBUG oslo_concurrency.lockutils [req-ae5a90ca-b359-4e22-b8b6-462b0f7d1b58 req-5af67daf-033d-46f8-baf9-d6c49cf0ba8a service nova] Releasing lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1202.092461] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-754c53f0-4f83-4fea-90eb-25bfc6eba922 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.099703] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144b4db0-e92a-4b24-a154-c6c8df261bbe {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.128356] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-760592f5-9da5-4482-bab6-be6cfeb4d6c3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.135141] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0613c7fd-1f21-4cf4-9fb8-8eed5cdc528b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.148085] env[62600]: DEBUG nova.compute.provider_tree [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1202.650808] env[62600]: DEBUG nova.scheduler.client.report [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1203.155637] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.118s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1203.156202] env[62600]: DEBUG nova.compute.manager [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Start building networks asynchronously for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1203.661167] env[62600]: DEBUG nova.compute.utils [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Using /dev/sd instead of None {{(pid=62600) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1203.662641] env[62600]: DEBUG nova.compute.manager [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Allocating IP information in the background. {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1203.662792] env[62600]: DEBUG nova.network.neutron [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] allocate_for_instance() {{(pid=62600) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1203.700569] env[62600]: DEBUG nova.policy [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03149ddca4f440b2b19d65fa53e7a8b2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3091650617964d398efef937402ebfdb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62600) authorize /opt/stack/nova/nova/policy.py:201}} [ 1203.934360] env[62600]: DEBUG nova.network.neutron [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Successfully created port: c4b3dc01-79ef-416f-a6a5-a0009c6e8648 {{(pid=62600) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1204.165890] env[62600]: DEBUG nova.compute.manager [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Start building block device mappings for instance. {{(pid=62600) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1205.177170] env[62600]: DEBUG nova.compute.manager [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Start spawning the instance on the hypervisor. {{(pid=62600) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1205.202110] env[62600]: DEBUG nova.virt.hardware [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T11:57:56Z,direct_url=,disk_format='vmdk',id=e7bd1d77-bfff-4684-9545-f6a1c69efe58,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ec39f2547f624ed0b240d084b949694a',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T11:57:57Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1205.202382] env[62600]: DEBUG nova.virt.hardware [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1205.202542] env[62600]: DEBUG nova.virt.hardware [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1205.202727] env[62600]: DEBUG nova.virt.hardware [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1205.202890] env[62600]: DEBUG nova.virt.hardware [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1205.203098] env[62600]: DEBUG nova.virt.hardware [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1205.203321] env[62600]: DEBUG nova.virt.hardware [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1205.203489] env[62600]: DEBUG nova.virt.hardware [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1205.203659] env[62600]: DEBUG nova.virt.hardware [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1205.203824] env[62600]: DEBUG nova.virt.hardware [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1205.203998] env[62600]: DEBUG nova.virt.hardware [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1205.204876] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b7667a-bcdc-403e-af01-05fe46093362 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.213178] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8248cae0-0812-4fbc-be83-c335758fcc32 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.299822] env[62600]: DEBUG nova.compute.manager [req-4f89a6d8-a7c4-43f1-9caa-f91d56a7662b req-a4a27039-1b56-4c58-9afe-ed5864ce2d36 service nova] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Received event network-vif-plugged-c4b3dc01-79ef-416f-a6a5-a0009c6e8648 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1205.300070] env[62600]: DEBUG oslo_concurrency.lockutils [req-4f89a6d8-a7c4-43f1-9caa-f91d56a7662b req-a4a27039-1b56-4c58-9afe-ed5864ce2d36 service nova] Acquiring lock "d34b306a-ba12-4923-b3e3-76db3bc1cd9e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1205.300338] env[62600]: DEBUG oslo_concurrency.lockutils [req-4f89a6d8-a7c4-43f1-9caa-f91d56a7662b req-a4a27039-1b56-4c58-9afe-ed5864ce2d36 service nova] Lock "d34b306a-ba12-4923-b3e3-76db3bc1cd9e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.300522] env[62600]: DEBUG oslo_concurrency.lockutils [req-4f89a6d8-a7c4-43f1-9caa-f91d56a7662b req-a4a27039-1b56-4c58-9afe-ed5864ce2d36 service nova] Lock "d34b306a-ba12-4923-b3e3-76db3bc1cd9e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1205.300698] env[62600]: DEBUG nova.compute.manager [req-4f89a6d8-a7c4-43f1-9caa-f91d56a7662b req-a4a27039-1b56-4c58-9afe-ed5864ce2d36 service nova] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] No waiting events found dispatching network-vif-plugged-c4b3dc01-79ef-416f-a6a5-a0009c6e8648 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1205.300870] env[62600]: WARNING nova.compute.manager [req-4f89a6d8-a7c4-43f1-9caa-f91d56a7662b req-a4a27039-1b56-4c58-9afe-ed5864ce2d36 service nova] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Received unexpected event network-vif-plugged-c4b3dc01-79ef-416f-a6a5-a0009c6e8648 for instance with vm_state building and task_state spawning. [ 1205.377808] env[62600]: DEBUG nova.network.neutron [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Successfully updated port: c4b3dc01-79ef-416f-a6a5-a0009c6e8648 {{(pid=62600) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1205.881098] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "refresh_cache-d34b306a-ba12-4923-b3e3-76db3bc1cd9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1205.881278] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquired lock "refresh_cache-d34b306a-ba12-4923-b3e3-76db3bc1cd9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1205.881434] env[62600]: DEBUG nova.network.neutron [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1206.411225] env[62600]: DEBUG nova.network.neutron [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Instance cache missing network info. {{(pid=62600) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1206.513877] env[62600]: DEBUG nova.network.neutron [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Updating instance_info_cache with network_info: [{"id": "c4b3dc01-79ef-416f-a6a5-a0009c6e8648", "address": "fa:16:3e:2a:5a:08", "network": {"id": "bcd55d3c-9d84-4e1d-b1fd-d6f1ac0cff02", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-771276928-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3091650617964d398efef937402ebfdb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4b3dc01-79", "ovs_interfaceid": "c4b3dc01-79ef-416f-a6a5-a0009c6e8648", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1207.016651] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Releasing lock "refresh_cache-d34b306a-ba12-4923-b3e3-76db3bc1cd9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1207.016976] env[62600]: DEBUG nova.compute.manager [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Instance network_info: |[{"id": "c4b3dc01-79ef-416f-a6a5-a0009c6e8648", "address": "fa:16:3e:2a:5a:08", "network": {"id": "bcd55d3c-9d84-4e1d-b1fd-d6f1ac0cff02", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-771276928-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3091650617964d398efef937402ebfdb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4b3dc01-79", "ovs_interfaceid": "c4b3dc01-79ef-416f-a6a5-a0009c6e8648", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62600) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1207.017473] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2a:5a:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a10c88d7-d13f-44fd-acee-7a734eb5f56a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c4b3dc01-79ef-416f-a6a5-a0009c6e8648', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1207.024872] env[62600]: DEBUG oslo.service.loopingcall [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1207.025111] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1207.025348] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-42d0d4cc-67af-4d64-aed7-d855f72bebd1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.045660] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1207.045660] env[62600]: value = "task-1223125" [ 1207.045660] env[62600]: _type = "Task" [ 1207.045660] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.053015] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1223125, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.326358] env[62600]: DEBUG nova.compute.manager [req-f1a7ff74-9541-4553-b767-724d3f9054b5 req-dc8a4b9a-0166-4037-8461-22cadcbf4167 service nova] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Received event network-changed-c4b3dc01-79ef-416f-a6a5-a0009c6e8648 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1207.326584] env[62600]: DEBUG nova.compute.manager [req-f1a7ff74-9541-4553-b767-724d3f9054b5 req-dc8a4b9a-0166-4037-8461-22cadcbf4167 service nova] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Refreshing instance network info cache due to event network-changed-c4b3dc01-79ef-416f-a6a5-a0009c6e8648. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1207.326914] env[62600]: DEBUG oslo_concurrency.lockutils [req-f1a7ff74-9541-4553-b767-724d3f9054b5 req-dc8a4b9a-0166-4037-8461-22cadcbf4167 service nova] Acquiring lock "refresh_cache-d34b306a-ba12-4923-b3e3-76db3bc1cd9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1207.326914] env[62600]: DEBUG oslo_concurrency.lockutils [req-f1a7ff74-9541-4553-b767-724d3f9054b5 req-dc8a4b9a-0166-4037-8461-22cadcbf4167 service nova] Acquired lock "refresh_cache-d34b306a-ba12-4923-b3e3-76db3bc1cd9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1207.327240] env[62600]: DEBUG nova.network.neutron [req-f1a7ff74-9541-4553-b767-724d3f9054b5 req-dc8a4b9a-0166-4037-8461-22cadcbf4167 service nova] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Refreshing network info cache for port c4b3dc01-79ef-416f-a6a5-a0009c6e8648 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1207.555188] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1223125, 'name': CreateVM_Task, 'duration_secs': 0.308658} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.555548] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1207.555981] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1207.556196] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1207.556524] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1207.556766] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e46685b-666a-403b-8070-9c26ef94ff73 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.561717] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1207.561717] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52572df7-5ad2-8aec-2c68-a1c53171c192" [ 1207.561717] env[62600]: _type = "Task" [ 1207.561717] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.569218] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52572df7-5ad2-8aec-2c68-a1c53171c192, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.995205] env[62600]: DEBUG nova.network.neutron [req-f1a7ff74-9541-4553-b767-724d3f9054b5 req-dc8a4b9a-0166-4037-8461-22cadcbf4167 service nova] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Updated VIF entry in instance network info cache for port c4b3dc01-79ef-416f-a6a5-a0009c6e8648. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1207.995578] env[62600]: DEBUG nova.network.neutron [req-f1a7ff74-9541-4553-b767-724d3f9054b5 req-dc8a4b9a-0166-4037-8461-22cadcbf4167 service nova] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Updating instance_info_cache with network_info: [{"id": "c4b3dc01-79ef-416f-a6a5-a0009c6e8648", "address": "fa:16:3e:2a:5a:08", "network": {"id": "bcd55d3c-9d84-4e1d-b1fd-d6f1ac0cff02", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-771276928-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3091650617964d398efef937402ebfdb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4b3dc01-79", "ovs_interfaceid": "c4b3dc01-79ef-416f-a6a5-a0009c6e8648", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1208.072819] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52572df7-5ad2-8aec-2c68-a1c53171c192, 'name': SearchDatastore_Task, 'duration_secs': 0.010344} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.073230] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1208.073557] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1208.073903] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1208.074114] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1208.074278] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1208.074532] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ed24151-b6cd-4d42-986a-477631758825 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.083272] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1208.083451] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1208.084139] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d82adc83-55f3-43c4-9ba6-02aab687cce0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.088965] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1208.088965] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5202c1ab-ca15-7675-5d64-fc72ecbb98ad" [ 1208.088965] env[62600]: _type = "Task" [ 1208.088965] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.096191] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5202c1ab-ca15-7675-5d64-fc72ecbb98ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.498210] env[62600]: DEBUG oslo_concurrency.lockutils [req-f1a7ff74-9541-4553-b767-724d3f9054b5 req-dc8a4b9a-0166-4037-8461-22cadcbf4167 service nova] Releasing lock "refresh_cache-d34b306a-ba12-4923-b3e3-76db3bc1cd9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1208.600158] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5202c1ab-ca15-7675-5d64-fc72ecbb98ad, 'name': SearchDatastore_Task, 'duration_secs': 0.007825} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.600985] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3097a11-221e-4b5f-b747-3fc393c44da7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.606373] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1208.606373] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521acab8-71bf-d2cc-ce10-e9b8f8635408" [ 1208.606373] env[62600]: _type = "Task" [ 1208.606373] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.613865] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521acab8-71bf-d2cc-ce10-e9b8f8635408, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.116672] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]521acab8-71bf-d2cc-ce10-e9b8f8635408, 'name': SearchDatastore_Task, 'duration_secs': 0.009301} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.116960] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1209.117257] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] d34b306a-ba12-4923-b3e3-76db3bc1cd9e/d34b306a-ba12-4923-b3e3-76db3bc1cd9e.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1209.117520] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fabd8a11-e206-400a-8407-2888512f8b99 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.125286] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1209.125286] env[62600]: value = "task-1223126" [ 1209.125286] env[62600]: _type = "Task" [ 1209.125286] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.132605] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223126, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.635430] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223126, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.415382} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.635839] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] d34b306a-ba12-4923-b3e3-76db3bc1cd9e/d34b306a-ba12-4923-b3e3-76db3bc1cd9e.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1209.635913] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Extending root virtual disk to 1048576 {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1209.636166] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5838bd81-7358-4c06-b5ad-a2913224e0d6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.642618] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1209.642618] env[62600]: value = "task-1223127" [ 1209.642618] env[62600]: _type = "Task" [ 1209.642618] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.649856] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223127, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.153437] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223127, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056673} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.153780] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Extended root virtual disk {{(pid=62600) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1210.154632] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a321ec9-f275-44ba-a86f-afe4d5645991 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.178505] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] d34b306a-ba12-4923-b3e3-76db3bc1cd9e/d34b306a-ba12-4923-b3e3-76db3bc1cd9e.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1210.178816] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-894e26c6-a49b-4e86-b38d-0cb59f644f76 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.199790] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1210.199790] env[62600]: value = "task-1223128" [ 1210.199790] env[62600]: _type = "Task" [ 1210.199790] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.208646] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223128, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.710175] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223128, 'name': ReconfigVM_Task, 'duration_secs': 0.274369} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.710506] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Reconfigured VM instance instance-0000006b to attach disk [datastore1] d34b306a-ba12-4923-b3e3-76db3bc1cd9e/d34b306a-ba12-4923-b3e3-76db3bc1cd9e.vmdk or device None with type sparse {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1210.711090] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2192b7a0-dbe6-4f40-93d0-edb9c2e5b5e0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.717356] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1210.717356] env[62600]: value = "task-1223129" [ 1210.717356] env[62600]: _type = "Task" [ 1210.717356] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.725894] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223129, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.228143] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223129, 'name': Rename_Task, 'duration_secs': 0.135827} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.228404] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1211.228630] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a561bfc-34f3-48e4-b445-7f8a99676a48 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.235069] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1211.235069] env[62600]: value = "task-1223130" [ 1211.235069] env[62600]: _type = "Task" [ 1211.235069] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.242626] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223130, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.744925] env[62600]: DEBUG oslo_vmware.api [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223130, 'name': PowerOnVM_Task, 'duration_secs': 0.404201} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.745408] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1211.745461] env[62600]: INFO nova.compute.manager [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Took 6.57 seconds to spawn the instance on the hypervisor. [ 1211.745609] env[62600]: DEBUG nova.compute.manager [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1211.746375] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb7db9f-2cac-4fed-a787-6ed5c4e2f893 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.262800] env[62600]: INFO nova.compute.manager [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Took 11.25 seconds to build instance. [ 1212.765382] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f8e3fc41-7b3f-4dd3-9c9c-6515f8128014 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lock "d34b306a-ba12-4923-b3e3-76db3bc1cd9e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.755s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1213.101430] env[62600]: INFO nova.compute.manager [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Rescuing [ 1213.101710] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "refresh_cache-d34b306a-ba12-4923-b3e3-76db3bc1cd9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1213.101865] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquired lock "refresh_cache-d34b306a-ba12-4923-b3e3-76db3bc1cd9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1213.102053] env[62600]: DEBUG nova.network.neutron [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1213.824214] env[62600]: DEBUG nova.network.neutron [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Updating instance_info_cache with network_info: [{"id": "c4b3dc01-79ef-416f-a6a5-a0009c6e8648", "address": "fa:16:3e:2a:5a:08", "network": {"id": "bcd55d3c-9d84-4e1d-b1fd-d6f1ac0cff02", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-771276928-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3091650617964d398efef937402ebfdb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4b3dc01-79", "ovs_interfaceid": "c4b3dc01-79ef-416f-a6a5-a0009c6e8648", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1214.327309] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Releasing lock "refresh_cache-d34b306a-ba12-4923-b3e3-76db3bc1cd9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1214.853579] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1214.853894] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95d5f406-cf24-4973-862d-a82e7ed38e05 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.861455] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1214.861455] env[62600]: value = "task-1223131" [ 1214.861455] env[62600]: _type = "Task" [ 1214.861455] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.869127] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223131, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.371272] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223131, 'name': PowerOffVM_Task, 'duration_secs': 0.189418} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.371542] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1215.372299] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b310454-3273-47ab-b254-0a59c12a5739 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.390959] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f53b49-95b6-4fe1-aebe-baba7cfe8bc4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.416400] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1215.416663] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-297fa824-d93a-416d-ba04-f984ef3e1f36 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.422691] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1215.422691] env[62600]: value = "task-1223132" [ 1215.422691] env[62600]: _type = "Task" [ 1215.422691] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.429894] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223132, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.933867] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] VM already powered off {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1215.934136] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Processing image e7bd1d77-bfff-4684-9545-f6a1c69efe58 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1215.934301] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1215.934459] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquired lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1215.934639] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1215.934885] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-36253a13-0eca-4dad-b642-bde880ee67eb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.943189] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1215.943377] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1215.944101] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee8b63f9-20be-4f3e-8a55-12fd5201a9a3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.949207] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1215.949207] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5262aa47-18af-087c-fd7e-83bfdca0720b" [ 1215.949207] env[62600]: _type = "Task" [ 1215.949207] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.957043] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5262aa47-18af-087c-fd7e-83bfdca0720b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.461054] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]5262aa47-18af-087c-fd7e-83bfdca0720b, 'name': SearchDatastore_Task, 'duration_secs': 0.008046} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.461054] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c786d32-c3a5-4e40-8861-0a642cbd11a3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.465659] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1216.465659] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52bf8f6e-d795-c2a3-88ac-74200bc59459" [ 1216.465659] env[62600]: _type = "Task" [ 1216.465659] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.473264] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52bf8f6e-d795-c2a3-88ac-74200bc59459, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.976127] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52bf8f6e-d795-c2a3-88ac-74200bc59459, 'name': SearchDatastore_Task, 'duration_secs': 0.009402} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.976540] env[62600]: DEBUG oslo_concurrency.lockutils [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Releasing lock "[datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1216.976670] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] d34b306a-ba12-4923-b3e3-76db3bc1cd9e/e7bd1d77-bfff-4684-9545-f6a1c69efe58-rescue.vmdk. {{(pid=62600) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1216.976923] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4db825ea-b01c-4666-9dcb-333b99e45fd3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.983673] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1216.983673] env[62600]: value = "task-1223133" [ 1216.983673] env[62600]: _type = "Task" [ 1216.983673] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.990893] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223133, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.494146] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223133, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.401331} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.494431] env[62600]: INFO nova.virt.vmwareapi.ds_util [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/e7bd1d77-bfff-4684-9545-f6a1c69efe58/e7bd1d77-bfff-4684-9545-f6a1c69efe58.vmdk to [datastore1] d34b306a-ba12-4923-b3e3-76db3bc1cd9e/e7bd1d77-bfff-4684-9545-f6a1c69efe58-rescue.vmdk. [ 1217.495209] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dc56abf-86f6-4bc7-9171-737f1c915056 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.518931] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] d34b306a-ba12-4923-b3e3-76db3bc1cd9e/e7bd1d77-bfff-4684-9545-f6a1c69efe58-rescue.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1217.519180] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-264148e4-5376-4453-a145-75245e6d8557 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.536236] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1217.536236] env[62600]: value = "task-1223134" [ 1217.536236] env[62600]: _type = "Task" [ 1217.536236] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.544219] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223134, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.046616] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223134, 'name': ReconfigVM_Task, 'duration_secs': 0.268462} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.047070] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Reconfigured VM instance instance-0000006b to attach disk [datastore1] d34b306a-ba12-4923-b3e3-76db3bc1cd9e/e7bd1d77-bfff-4684-9545-f6a1c69efe58-rescue.vmdk or device None with type thin {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1218.047683] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c4b923-e562-4667-ba39-0868cb166d3f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.071036] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-940dd4f6-ac8f-46ce-8dbe-d39152aca499 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.086874] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1218.086874] env[62600]: value = "task-1223135" [ 1218.086874] env[62600]: _type = "Task" [ 1218.086874] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.094171] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223135, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.596435] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223135, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.728186] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.096853] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223135, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.597086] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223135, 'name': ReconfigVM_Task, 'duration_secs': 1.171159} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.597417] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1219.597610] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2ec71089-ec5b-488c-a4e0-c022ff3f9839 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.603688] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1219.603688] env[62600]: value = "task-1223136" [ 1219.603688] env[62600]: _type = "Task" [ 1219.603688] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.610651] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223136, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.727471] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1219.727741] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.113591] env[62600]: DEBUG oslo_vmware.api [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223136, 'name': PowerOnVM_Task, 'duration_secs': 0.352829} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.115344] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1220.117365] env[62600]: DEBUG nova.compute.manager [None req-f6992b08-0a6d-499f-a6c0-c8e84f23c69a tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1220.118172] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ded57f0-dec2-40b9-bba3-164f34093925 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.231629] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1220.231860] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1220.232038] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1220.232200] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62600) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1220.233094] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56368024-2e70-4a5d-a796-2af306ce6b3a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.241022] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ef3a1a7-5283-4734-a79e-e0d3dca6ad19 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.253842] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29c5eddb-3b24-41c1-b1f7-d6cc328e26a8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.259638] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a06e20-0a0a-487b-bb1d-e559cdaa9cf8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.288389] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180852MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=62600) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1220.288533] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1220.288715] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1221.023072] env[62600]: INFO nova.compute.manager [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Unrescuing [ 1221.023072] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "refresh_cache-d34b306a-ba12-4923-b3e3-76db3bc1cd9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1221.023072] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquired lock "refresh_cache-d34b306a-ba12-4923-b3e3-76db3bc1cd9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1221.023072] env[62600]: DEBUG nova.network.neutron [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1221.315136] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance e1a6e69d-09d8-44f0-b3a6-405ab3820c07 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1221.315431] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 00648d57-68ae-438b-af37-7e50d6cdcedd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1221.315431] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance d34b306a-ba12-4923-b3e3-76db3bc1cd9e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1221.315606] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1221.315746] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1221.359860] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f274de01-9e3b-4543-aae3-cc2ddd75e35b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.367177] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bcd0012-611d-42f4-95b2-73b24ccc779d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.397087] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea89e2a-bf39-45a0-83d2-feafd22bc738 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.403965] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30418b56-6187-4b0c-88c1-9a2be5da1c81 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.417088] env[62600]: DEBUG nova.compute.provider_tree [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1221.727840] env[62600]: DEBUG nova.network.neutron [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Updating instance_info_cache with network_info: [{"id": "c4b3dc01-79ef-416f-a6a5-a0009c6e8648", "address": "fa:16:3e:2a:5a:08", "network": {"id": "bcd55d3c-9d84-4e1d-b1fd-d6f1ac0cff02", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-771276928-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "3091650617964d398efef937402ebfdb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4b3dc01-79", "ovs_interfaceid": "c4b3dc01-79ef-416f-a6a5-a0009c6e8648", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1221.920116] env[62600]: DEBUG nova.scheduler.client.report [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1222.230743] env[62600]: DEBUG oslo_concurrency.lockutils [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Releasing lock "refresh_cache-d34b306a-ba12-4923-b3e3-76db3bc1cd9e" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1222.231506] env[62600]: DEBUG nova.objects.instance [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lazy-loading 'flavor' on Instance uuid d34b306a-ba12-4923-b3e3-76db3bc1cd9e {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1222.425039] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62600) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1222.425501] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.136s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1222.737513] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb26cff-07e8-4a8f-aafb-111d820da37c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.759036] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1222.759036] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-05db4006-91e9-46e3-a0b5-1d9aece1c5ed {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.765637] env[62600]: DEBUG oslo_vmware.api [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1222.765637] env[62600]: value = "task-1223137" [ 1222.765637] env[62600]: _type = "Task" [ 1222.765637] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.773736] env[62600]: DEBUG oslo_vmware.api [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223137, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.275584] env[62600]: DEBUG oslo_vmware.api [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223137, 'name': PowerOffVM_Task, 'duration_secs': 0.218132} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.275850] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1223.280986] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Reconfiguring VM instance instance-0000006b to detach disk 2001 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1223.281222] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e8a1c3c-adc6-45de-98a8-4fa49bf379bd {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.300107] env[62600]: DEBUG oslo_vmware.api [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1223.300107] env[62600]: value = "task-1223138" [ 1223.300107] env[62600]: _type = "Task" [ 1223.300107] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.307262] env[62600]: DEBUG oslo_vmware.api [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223138, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.426205] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1223.426618] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1223.426719] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1223.723856] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1223.727490] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1223.727642] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Starting heal instance info cache {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1223.810323] env[62600]: DEBUG oslo_vmware.api [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223138, 'name': ReconfigVM_Task, 'duration_secs': 0.196824} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.810594] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Reconfigured VM instance instance-0000006b to detach disk 2001 {{(pid=62600) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1223.810792] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1223.811048] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d72c2ab6-5d18-4fef-9ce9-d4634de1c615 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.817246] env[62600]: DEBUG oslo_vmware.api [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1223.817246] env[62600]: value = "task-1223139" [ 1223.817246] env[62600]: _type = "Task" [ 1223.817246] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.824134] env[62600]: DEBUG oslo_vmware.api [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223139, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.230492] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Didn't find any instances for network info cache update. {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1224.327268] env[62600]: DEBUG oslo_vmware.api [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223139, 'name': PowerOnVM_Task, 'duration_secs': 0.344264} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.327573] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1224.327835] env[62600]: DEBUG nova.compute.manager [None req-fef94280-9423-4d92-a74c-f0e44554376d tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1224.328660] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-057b0ed5-8aa7-4566-9319-ef8369a5ac80 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.225864] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1225.653613] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "d34b306a-ba12-4923-b3e3-76db3bc1cd9e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1225.653893] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lock "d34b306a-ba12-4923-b3e3-76db3bc1cd9e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1225.654202] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "d34b306a-ba12-4923-b3e3-76db3bc1cd9e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1225.654411] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lock "d34b306a-ba12-4923-b3e3-76db3bc1cd9e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1225.654588] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lock "d34b306a-ba12-4923-b3e3-76db3bc1cd9e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1225.656822] env[62600]: INFO nova.compute.manager [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Terminating instance [ 1225.659012] env[62600]: DEBUG nova.compute.manager [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1225.659222] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1225.660052] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4519b1a8-fb33-4c12-9823-0d7019954ab1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.668068] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1225.668301] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e857ddc6-4679-4b53-b45e-648aeddcbf55 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.674740] env[62600]: DEBUG oslo_vmware.api [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1225.674740] env[62600]: value = "task-1223140" [ 1225.674740] env[62600]: _type = "Task" [ 1225.674740] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.681983] env[62600]: DEBUG oslo_vmware.api [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223140, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.184471] env[62600]: DEBUG oslo_vmware.api [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223140, 'name': PowerOffVM_Task, 'duration_secs': 0.171933} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.184760] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1226.184904] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1226.185228] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b43fba6b-c2e6-4d07-8a0a-6e04ce1f2c42 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.243847] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1226.244103] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Deleting contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1226.244284] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Deleting the datastore file [datastore1] d34b306a-ba12-4923-b3e3-76db3bc1cd9e {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1226.244585] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d430fe27-47dc-4a75-8a97-e56d7a929ed8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.251258] env[62600]: DEBUG oslo_vmware.api [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1226.251258] env[62600]: value = "task-1223142" [ 1226.251258] env[62600]: _type = "Task" [ 1226.251258] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.260905] env[62600]: DEBUG oslo_vmware.api [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223142, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.727720] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1226.727927] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62600) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1226.760832] env[62600]: DEBUG oslo_vmware.api [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223142, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133577} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.761091] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1226.761282] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Deleted contents of the VM from datastore datastore1 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1226.761462] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1226.761656] env[62600]: INFO nova.compute.manager [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1226.761874] env[62600]: DEBUG oslo.service.loopingcall [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1226.762075] env[62600]: DEBUG nova.compute.manager [-] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1226.762168] env[62600]: DEBUG nova.network.neutron [-] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1226.998487] env[62600]: DEBUG nova.compute.manager [req-f108ce94-bdcd-496f-afc9-a4033865afbd req-7b6c3164-e874-4170-bfb7-2440c2584d07 service nova] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Received event network-vif-deleted-c4b3dc01-79ef-416f-a6a5-a0009c6e8648 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1226.998667] env[62600]: INFO nova.compute.manager [req-f108ce94-bdcd-496f-afc9-a4033865afbd req-7b6c3164-e874-4170-bfb7-2440c2584d07 service nova] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Neutron deleted interface c4b3dc01-79ef-416f-a6a5-a0009c6e8648; detaching it from the instance and deleting it from the info cache [ 1226.998804] env[62600]: DEBUG nova.network.neutron [req-f108ce94-bdcd-496f-afc9-a4033865afbd req-7b6c3164-e874-4170-bfb7-2440c2584d07 service nova] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1227.479377] env[62600]: DEBUG nova.network.neutron [-] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1227.501280] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-00d1930d-1a0e-4185-92a3-d279fbfebe0b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.510561] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262d57f1-a190-4a46-9c90-9eaec647c00d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.532909] env[62600]: DEBUG nova.compute.manager [req-f108ce94-bdcd-496f-afc9-a4033865afbd req-7b6c3164-e874-4170-bfb7-2440c2584d07 service nova] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Detach interface failed, port_id=c4b3dc01-79ef-416f-a6a5-a0009c6e8648, reason: Instance d34b306a-ba12-4923-b3e3-76db3bc1cd9e could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1227.981485] env[62600]: INFO nova.compute.manager [-] [instance: d34b306a-ba12-4923-b3e3-76db3bc1cd9e] Took 1.22 seconds to deallocate network for instance. [ 1228.487768] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1228.488078] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1228.488308] env[62600]: DEBUG nova.objects.instance [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lazy-loading 'resources' on Instance uuid d34b306a-ba12-4923-b3e3-76db3bc1cd9e {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1229.040185] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28bc0803-2568-433c-9a4b-b49b703bf15a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.047636] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d7c1fca-c4fb-4c5b-b92f-2bcbef604b3a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.077408] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c98f284-e22c-44bb-be8f-ca5e8da182ed {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.084462] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-895d8999-2ab0-4151-b8e0-a70e83534e14 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.097069] env[62600]: DEBUG nova.compute.provider_tree [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1229.600544] env[62600]: DEBUG nova.scheduler.client.report [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1230.105549] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.617s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1230.127876] env[62600]: INFO nova.scheduler.client.report [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Deleted allocations for instance d34b306a-ba12-4923-b3e3-76db3bc1cd9e [ 1230.636024] env[62600]: DEBUG oslo_concurrency.lockutils [None req-e88fa6df-a84f-42ba-9060-7acc97743855 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lock "d34b306a-ba12-4923-b3e3-76db3bc1cd9e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.982s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1231.117019] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "e1a6e69d-09d8-44f0-b3a6-405ab3820c07" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1231.117334] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lock "e1a6e69d-09d8-44f0-b3a6-405ab3820c07" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1231.117617] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "e1a6e69d-09d8-44f0-b3a6-405ab3820c07-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1231.117855] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lock "e1a6e69d-09d8-44f0-b3a6-405ab3820c07-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1231.118049] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lock "e1a6e69d-09d8-44f0-b3a6-405ab3820c07-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1231.120185] env[62600]: INFO nova.compute.manager [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Terminating instance [ 1231.121987] env[62600]: DEBUG nova.compute.manager [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1231.122212] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1231.123071] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9aa5fd6-1d76-4a52-a18d-d1563bb99ab1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.131478] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1231.131710] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d58bba5d-0cf4-44c4-a86e-3620a4a9914a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.138106] env[62600]: DEBUG oslo_vmware.api [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1231.138106] env[62600]: value = "task-1223143" [ 1231.138106] env[62600]: _type = "Task" [ 1231.138106] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.145140] env[62600]: DEBUG oslo_vmware.api [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223143, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.648587] env[62600]: DEBUG oslo_vmware.api [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223143, 'name': PowerOffVM_Task, 'duration_secs': 0.216672} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.648985] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1231.649062] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1231.649318] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4f6f2f31-0e67-44ff-b088-872568dbe4f3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.711236] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1231.711507] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1231.711644] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Deleting the datastore file [datastore2] e1a6e69d-09d8-44f0-b3a6-405ab3820c07 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1231.711919] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bbc6565f-e7d7-490c-abad-34b0dff1a767 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.717934] env[62600]: DEBUG oslo_vmware.api [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for the task: (returnval){ [ 1231.717934] env[62600]: value = "task-1223145" [ 1231.717934] env[62600]: _type = "Task" [ 1231.717934] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.726060] env[62600]: DEBUG oslo_vmware.api [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223145, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.229624] env[62600]: DEBUG oslo_vmware.api [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Task: {'id': task-1223145, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.185207} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.229800] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1232.229944] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1232.230139] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1232.230320] env[62600]: INFO nova.compute.manager [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1232.230570] env[62600]: DEBUG oslo.service.loopingcall [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1232.230765] env[62600]: DEBUG nova.compute.manager [-] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1232.230859] env[62600]: DEBUG nova.network.neutron [-] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1232.516568] env[62600]: DEBUG nova.compute.manager [req-8f3893da-1b51-4fcb-a172-a806fc3a19a8 req-ac03a0c5-dc6c-45a4-aa1a-4ce25b3a01ea service nova] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Received event network-vif-deleted-4de0ebd7-1530-47fd-8653-72b841a6d4d1 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1232.516819] env[62600]: INFO nova.compute.manager [req-8f3893da-1b51-4fcb-a172-a806fc3a19a8 req-ac03a0c5-dc6c-45a4-aa1a-4ce25b3a01ea service nova] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Neutron deleted interface 4de0ebd7-1530-47fd-8653-72b841a6d4d1; detaching it from the instance and deleting it from the info cache [ 1232.516974] env[62600]: DEBUG nova.network.neutron [req-8f3893da-1b51-4fcb-a172-a806fc3a19a8 req-ac03a0c5-dc6c-45a4-aa1a-4ce25b3a01ea service nova] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1232.995039] env[62600]: DEBUG nova.network.neutron [-] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1233.019583] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-76b30360-0e8c-434a-bb5d-054d8b0ff35a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.029962] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-053c68f3-a21f-4197-81fe-3548058e7784 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.054932] env[62600]: DEBUG nova.compute.manager [req-8f3893da-1b51-4fcb-a172-a806fc3a19a8 req-ac03a0c5-dc6c-45a4-aa1a-4ce25b3a01ea service nova] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Detach interface failed, port_id=4de0ebd7-1530-47fd-8653-72b841a6d4d1, reason: Instance e1a6e69d-09d8-44f0-b3a6-405ab3820c07 could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1233.497684] env[62600]: INFO nova.compute.manager [-] [instance: e1a6e69d-09d8-44f0-b3a6-405ab3820c07] Took 1.27 seconds to deallocate network for instance. [ 1234.004260] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1234.004558] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.004789] env[62600]: DEBUG nova.objects.instance [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lazy-loading 'resources' on Instance uuid e1a6e69d-09d8-44f0-b3a6-405ab3820c07 {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1234.551013] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a09d626-a3d8-4692-97c0-819269d4ba4d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.557400] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a87efa85-cc49-4bdd-8437-2359c9368d7a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.587884] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da0a8a0-97c1-4449-9312-989f9777d085 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.594702] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b354cf98-174e-439e-b9a1-ec7b86d4ea77 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.607405] env[62600]: DEBUG nova.compute.provider_tree [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1235.111063] env[62600]: DEBUG nova.scheduler.client.report [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1235.616712] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.612s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1235.635287] env[62600]: INFO nova.scheduler.client.report [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Deleted allocations for instance e1a6e69d-09d8-44f0-b3a6-405ab3820c07 [ 1236.142864] env[62600]: DEBUG oslo_concurrency.lockutils [None req-4d686354-e1bc-47e0-bbb3-d28dcd913359 tempest-ServerRescueTestJSON-1048724030 tempest-ServerRescueTestJSON-1048724030-project-member] Lock "e1a6e69d-09d8-44f0-b3a6-405ab3820c07" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.025s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1238.267667] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "00648d57-68ae-438b-af37-7e50d6cdcedd" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1238.268123] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "00648d57-68ae-438b-af37-7e50d6cdcedd" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1238.268123] env[62600]: INFO nova.compute.manager [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Shelving [ 1238.775987] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1238.776270] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-355bbeb5-ddb5-4159-a646-35f6efa39788 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.782986] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1238.782986] env[62600]: value = "task-1223146" [ 1238.782986] env[62600]: _type = "Task" [ 1238.782986] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.791682] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223146, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.292754] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223146, 'name': PowerOffVM_Task, 'duration_secs': 0.165028} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.293154] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1239.293756] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe5900d0-9e8d-4ce3-99cc-64d6a7e2f5d5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.312336] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-372be5e5-bb85-4b34-9171-2db8611fbca2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.822232] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Creating Snapshot of the VM instance {{(pid=62600) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1239.822590] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1934cbe1-8a8b-4efe-9fcd-968a825783b7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.830537] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1239.830537] env[62600]: value = "task-1223147" [ 1239.830537] env[62600]: _type = "Task" [ 1239.830537] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.838870] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223147, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.340620] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223147, 'name': CreateSnapshot_Task, 'duration_secs': 0.428185} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.341042] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Created Snapshot of the VM instance {{(pid=62600) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1240.341624] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44ee6f87-81a7-4674-a440-65080430a11e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.858755] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Creating linked-clone VM from snapshot {{(pid=62600) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1240.859068] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6cbf7fb2-fa1e-4f46-8c58-4d3e789f1c32 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.868573] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1240.868573] env[62600]: value = "task-1223148" [ 1240.868573] env[62600]: _type = "Task" [ 1240.868573] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.876218] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223148, 'name': CloneVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.378858] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223148, 'name': CloneVM_Task} progress is 94%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.879564] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223148, 'name': CloneVM_Task, 'duration_secs': 0.898167} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.879830] env[62600]: INFO nova.virt.vmwareapi.vmops [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Created linked-clone VM from snapshot [ 1241.880555] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc7e0ad5-a798-4797-b785-e3818f135388 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.887138] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Uploading image 27540ab1-93b4-4ac2-bd01-e836c555bff5 {{(pid=62600) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1241.910211] env[62600]: DEBUG oslo_vmware.rw_handles [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1241.910211] env[62600]: value = "vm-264387" [ 1241.910211] env[62600]: _type = "VirtualMachine" [ 1241.910211] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1241.910458] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f8e1cf6f-f633-47d2-8659-e00b19c6ea1c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.916781] env[62600]: DEBUG oslo_vmware.rw_handles [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lease: (returnval){ [ 1241.916781] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ed4734-f4d8-e267-f73f-fd9c54316d6f" [ 1241.916781] env[62600]: _type = "HttpNfcLease" [ 1241.916781] env[62600]: } obtained for exporting VM: (result){ [ 1241.916781] env[62600]: value = "vm-264387" [ 1241.916781] env[62600]: _type = "VirtualMachine" [ 1241.916781] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1241.917048] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the lease: (returnval){ [ 1241.917048] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ed4734-f4d8-e267-f73f-fd9c54316d6f" [ 1241.917048] env[62600]: _type = "HttpNfcLease" [ 1241.917048] env[62600]: } to be ready. {{(pid=62600) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1241.922839] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1241.922839] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ed4734-f4d8-e267-f73f-fd9c54316d6f" [ 1241.922839] env[62600]: _type = "HttpNfcLease" [ 1241.922839] env[62600]: } is initializing. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1242.424799] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1242.424799] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ed4734-f4d8-e267-f73f-fd9c54316d6f" [ 1242.424799] env[62600]: _type = "HttpNfcLease" [ 1242.424799] env[62600]: } is ready. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1242.425322] env[62600]: DEBUG oslo_vmware.rw_handles [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1242.425322] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52ed4734-f4d8-e267-f73f-fd9c54316d6f" [ 1242.425322] env[62600]: _type = "HttpNfcLease" [ 1242.425322] env[62600]: }. {{(pid=62600) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1242.425823] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5021d1b3-4e04-4fba-8dbf-50952853d5b1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.432848] env[62600]: DEBUG oslo_vmware.rw_handles [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52883613-2820-3860-4fa1-9ffab39c4f52/disk-0.vmdk from lease info. {{(pid=62600) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1242.433037] env[62600]: DEBUG oslo_vmware.rw_handles [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52883613-2820-3860-4fa1-9ffab39c4f52/disk-0.vmdk for reading. {{(pid=62600) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1242.521849] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-62f129ab-9101-4074-947b-a721a5a45e05 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.843642] env[62600]: DEBUG oslo_vmware.rw_handles [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52883613-2820-3860-4fa1-9ffab39c4f52/disk-0.vmdk. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1249.844579] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9894f634-41d0-4f12-850c-4db04ecb1c43 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.850923] env[62600]: DEBUG oslo_vmware.rw_handles [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52883613-2820-3860-4fa1-9ffab39c4f52/disk-0.vmdk is in state: ready. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1249.851109] env[62600]: ERROR oslo_vmware.rw_handles [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52883613-2820-3860-4fa1-9ffab39c4f52/disk-0.vmdk due to incomplete transfer. [ 1249.851321] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1daa654f-bce7-446e-a276-1fdc94ab53fb {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.857828] env[62600]: DEBUG oslo_vmware.rw_handles [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52883613-2820-3860-4fa1-9ffab39c4f52/disk-0.vmdk. {{(pid=62600) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1249.858032] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Uploaded image 27540ab1-93b4-4ac2-bd01-e836c555bff5 to the Glance image server {{(pid=62600) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1249.860350] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Destroying the VM {{(pid=62600) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1249.860576] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-5eb05e38-25f9-400d-bcf8-579e7c706b6a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.866060] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1249.866060] env[62600]: value = "task-1223150" [ 1249.866060] env[62600]: _type = "Task" [ 1249.866060] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.873166] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223150, 'name': Destroy_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.375618] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223150, 'name': Destroy_Task, 'duration_secs': 0.31774} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.375885] env[62600]: INFO nova.virt.vmwareapi.vm_util [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Destroyed the VM [ 1250.376141] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Deleting Snapshot of the VM instance {{(pid=62600) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1250.376411] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-2bd1a274-af67-468d-a489-f1eeb881e6ae {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.382819] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1250.382819] env[62600]: value = "task-1223151" [ 1250.382819] env[62600]: _type = "Task" [ 1250.382819] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.390140] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223151, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.893123] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223151, 'name': RemoveSnapshot_Task, 'duration_secs': 0.351678} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.893505] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Deleted Snapshot of the VM instance {{(pid=62600) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1250.893730] env[62600]: DEBUG nova.compute.manager [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1250.894476] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b3b788-9650-41de-be05-1895f2ffe408 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.406670] env[62600]: INFO nova.compute.manager [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Shelve offloading [ 1251.408338] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1251.408588] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f1eef994-b794-490d-972d-70e4e2762ad0 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.416417] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1251.416417] env[62600]: value = "task-1223152" [ 1251.416417] env[62600]: _type = "Task" [ 1251.416417] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.423961] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223152, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.926302] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] VM already powered off {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1251.926691] env[62600]: DEBUG nova.compute.manager [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1251.927173] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cacbd8d-5e6d-44bc-aea7-6eab1ed8660a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.932424] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1251.932592] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquired lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1251.932763] env[62600]: DEBUG nova.network.neutron [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1252.635987] env[62600]: DEBUG nova.network.neutron [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Updating instance_info_cache with network_info: [{"id": "9c4c9eaa-2ecd-4598-bed6-f746cbe53f80", "address": "fa:16:3e:6b:09:8c", "network": {"id": "dc21424f-b891-428d-a74e-76ab78d757a8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1428037886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "952323d62f034126a9b46750d03fe6ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c4c9eaa-2e", "ovs_interfaceid": "9c4c9eaa-2ecd-4598-bed6-f746cbe53f80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1253.138793] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Releasing lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1253.342794] env[62600]: DEBUG nova.compute.manager [req-288eb6f6-44cc-419c-ae7d-acb8adb9e1d0 req-17feacc1-fd1c-4f13-86f3-5bc7747e158a service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Received event network-vif-unplugged-9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1253.343111] env[62600]: DEBUG oslo_concurrency.lockutils [req-288eb6f6-44cc-419c-ae7d-acb8adb9e1d0 req-17feacc1-fd1c-4f13-86f3-5bc7747e158a service nova] Acquiring lock "00648d57-68ae-438b-af37-7e50d6cdcedd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1253.343377] env[62600]: DEBUG oslo_concurrency.lockutils [req-288eb6f6-44cc-419c-ae7d-acb8adb9e1d0 req-17feacc1-fd1c-4f13-86f3-5bc7747e158a service nova] Lock "00648d57-68ae-438b-af37-7e50d6cdcedd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1253.343603] env[62600]: DEBUG oslo_concurrency.lockutils [req-288eb6f6-44cc-419c-ae7d-acb8adb9e1d0 req-17feacc1-fd1c-4f13-86f3-5bc7747e158a service nova] Lock "00648d57-68ae-438b-af37-7e50d6cdcedd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1253.343803] env[62600]: DEBUG nova.compute.manager [req-288eb6f6-44cc-419c-ae7d-acb8adb9e1d0 req-17feacc1-fd1c-4f13-86f3-5bc7747e158a service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] No waiting events found dispatching network-vif-unplugged-9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1253.344024] env[62600]: WARNING nova.compute.manager [req-288eb6f6-44cc-419c-ae7d-acb8adb9e1d0 req-17feacc1-fd1c-4f13-86f3-5bc7747e158a service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Received unexpected event network-vif-unplugged-9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 for instance with vm_state shelved and task_state shelving_offloading. [ 1253.431282] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1253.432191] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8046be45-e83d-458a-a1fc-1d890193c0aa {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.439828] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1253.440084] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d108c772-f6ac-44c8-a95f-5cb64ff585ef {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.515415] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1253.515668] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1253.515859] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Deleting the datastore file [datastore2] 00648d57-68ae-438b-af37-7e50d6cdcedd {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1253.516136] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3a09c671-3d74-4e33-9567-0b5fdd5c8d2a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.523452] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1253.523452] env[62600]: value = "task-1223154" [ 1253.523452] env[62600]: _type = "Task" [ 1253.523452] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.530865] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223154, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.033260] env[62600]: DEBUG oslo_vmware.api [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223154, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136374} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.033551] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1254.033740] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1254.033919] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1254.056122] env[62600]: INFO nova.scheduler.client.report [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Deleted allocations for instance 00648d57-68ae-438b-af37-7e50d6cdcedd [ 1254.562037] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1254.562355] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1254.562569] env[62600]: DEBUG nova.objects.instance [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lazy-loading 'resources' on Instance uuid 00648d57-68ae-438b-af37-7e50d6cdcedd {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1255.065689] env[62600]: DEBUG nova.objects.instance [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lazy-loading 'numa_topology' on Instance uuid 00648d57-68ae-438b-af37-7e50d6cdcedd {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1255.367348] env[62600]: DEBUG nova.compute.manager [req-10978c4a-b4f3-49a5-a08b-40ab42f10c19 req-d8429173-55f2-43bb-b9b9-f357636193b9 service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Received event network-changed-9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1255.367520] env[62600]: DEBUG nova.compute.manager [req-10978c4a-b4f3-49a5-a08b-40ab42f10c19 req-d8429173-55f2-43bb-b9b9-f357636193b9 service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Refreshing instance network info cache due to event network-changed-9c4c9eaa-2ecd-4598-bed6-f746cbe53f80. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1255.367813] env[62600]: DEBUG oslo_concurrency.lockutils [req-10978c4a-b4f3-49a5-a08b-40ab42f10c19 req-d8429173-55f2-43bb-b9b9-f357636193b9 service nova] Acquiring lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1255.367996] env[62600]: DEBUG oslo_concurrency.lockutils [req-10978c4a-b4f3-49a5-a08b-40ab42f10c19 req-d8429173-55f2-43bb-b9b9-f357636193b9 service nova] Acquired lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1255.368193] env[62600]: DEBUG nova.network.neutron [req-10978c4a-b4f3-49a5-a08b-40ab42f10c19 req-d8429173-55f2-43bb-b9b9-f357636193b9 service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Refreshing network info cache for port 9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1255.568711] env[62600]: DEBUG nova.objects.base [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Object Instance<00648d57-68ae-438b-af37-7e50d6cdcedd> lazy-loaded attributes: resources,numa_topology {{(pid=62600) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1255.582950] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6bbe413-ae2e-4b6f-b9ec-034463750099 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.591203] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-142a8479-c6f5-4e9b-b712-771704ff036b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.620832] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7cbed62-cdb7-47d3-9aac-4620b18f28f3 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.628379] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101e4a4b-4f30-4414-809a-0cbaf807fa12 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.641648] env[62600]: DEBUG nova.compute.provider_tree [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1256.062467] env[62600]: DEBUG nova.network.neutron [req-10978c4a-b4f3-49a5-a08b-40ab42f10c19 req-d8429173-55f2-43bb-b9b9-f357636193b9 service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Updated VIF entry in instance network info cache for port 9c4c9eaa-2ecd-4598-bed6-f746cbe53f80. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1256.062838] env[62600]: DEBUG nova.network.neutron [req-10978c4a-b4f3-49a5-a08b-40ab42f10c19 req-d8429173-55f2-43bb-b9b9-f357636193b9 service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Updating instance_info_cache with network_info: [{"id": "9c4c9eaa-2ecd-4598-bed6-f746cbe53f80", "address": "fa:16:3e:6b:09:8c", "network": {"id": "dc21424f-b891-428d-a74e-76ab78d757a8", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1428037886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "952323d62f034126a9b46750d03fe6ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap9c4c9eaa-2e", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1256.144311] env[62600]: DEBUG nova.scheduler.client.report [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1256.565916] env[62600]: DEBUG oslo_concurrency.lockutils [req-10978c4a-b4f3-49a5-a08b-40ab42f10c19 req-d8429173-55f2-43bb-b9b9-f357636193b9 service nova] Releasing lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1256.649100] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.087s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1256.699077] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "00648d57-68ae-438b-af37-7e50d6cdcedd" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1257.157059] env[62600]: DEBUG oslo_concurrency.lockutils [None req-27d99b28-ff9c-42d8-a8c0-a05418945ee1 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "00648d57-68ae-438b-af37-7e50d6cdcedd" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 18.889s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1257.157909] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "00648d57-68ae-438b-af37-7e50d6cdcedd" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.459s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1257.158113] env[62600]: INFO nova.compute.manager [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Unshelving [ 1258.179529] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1258.179807] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1258.180029] env[62600]: DEBUG nova.objects.instance [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lazy-loading 'pci_requests' on Instance uuid 00648d57-68ae-438b-af37-7e50d6cdcedd {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1258.683926] env[62600]: DEBUG nova.objects.instance [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lazy-loading 'numa_topology' on Instance uuid 00648d57-68ae-438b-af37-7e50d6cdcedd {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1259.186365] env[62600]: INFO nova.compute.claims [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1260.226254] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c91e7456-eb1b-4e4b-81ed-82767c02a850 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.233560] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5d75dc9-e938-424d-8a2e-f3bddc718782 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.263330] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af6ef364-b321-4527-81af-15875145a1ad {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.269890] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b45f0415-f233-4052-8984-241e6f48d513 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.282173] env[62600]: DEBUG nova.compute.provider_tree [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1260.785111] env[62600]: DEBUG nova.scheduler.client.report [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1261.290463] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.110s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1261.317838] env[62600]: INFO nova.network.neutron [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Updating port 9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1262.697943] env[62600]: DEBUG nova.compute.manager [req-ffab405e-62ed-4cf4-a173-4dc40694ce16 req-74deedfd-75e9-496f-a9a0-658ba14fb52a service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Received event network-vif-plugged-9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1262.698273] env[62600]: DEBUG oslo_concurrency.lockutils [req-ffab405e-62ed-4cf4-a173-4dc40694ce16 req-74deedfd-75e9-496f-a9a0-658ba14fb52a service nova] Acquiring lock "00648d57-68ae-438b-af37-7e50d6cdcedd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1262.698409] env[62600]: DEBUG oslo_concurrency.lockutils [req-ffab405e-62ed-4cf4-a173-4dc40694ce16 req-74deedfd-75e9-496f-a9a0-658ba14fb52a service nova] Lock "00648d57-68ae-438b-af37-7e50d6cdcedd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1262.698576] env[62600]: DEBUG oslo_concurrency.lockutils [req-ffab405e-62ed-4cf4-a173-4dc40694ce16 req-74deedfd-75e9-496f-a9a0-658ba14fb52a service nova] Lock "00648d57-68ae-438b-af37-7e50d6cdcedd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1262.698751] env[62600]: DEBUG nova.compute.manager [req-ffab405e-62ed-4cf4-a173-4dc40694ce16 req-74deedfd-75e9-496f-a9a0-658ba14fb52a service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] No waiting events found dispatching network-vif-plugged-9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 {{(pid=62600) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1262.698924] env[62600]: WARNING nova.compute.manager [req-ffab405e-62ed-4cf4-a173-4dc40694ce16 req-74deedfd-75e9-496f-a9a0-658ba14fb52a service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Received unexpected event network-vif-plugged-9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 for instance with vm_state shelved_offloaded and task_state spawning. [ 1262.786186] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1262.786307] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquired lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1262.786507] env[62600]: DEBUG nova.network.neutron [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Building network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1263.532525] env[62600]: DEBUG nova.network.neutron [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Updating instance_info_cache with network_info: [{"id": "9c4c9eaa-2ecd-4598-bed6-f746cbe53f80", "address": "fa:16:3e:6b:09:8c", "network": {"id": "dc21424f-b891-428d-a74e-76ab78d757a8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1428037886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "952323d62f034126a9b46750d03fe6ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c4c9eaa-2e", "ovs_interfaceid": "9c4c9eaa-2ecd-4598-bed6-f746cbe53f80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1264.035684] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Releasing lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1264.060904] env[62600]: DEBUG nova.virt.hardware [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T11:58:13Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='18ff0bb54db19551ab7b60b09d989c98',container_format='bare',created_at=2024-10-09T12:11:25Z,direct_url=,disk_format='vmdk',id=27540ab1-93b4-4ac2-bd01-e836c555bff5,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-208775560-shelved',owner='952323d62f034126a9b46750d03fe6ea',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2024-10-09T12:11:37Z,virtual_size=,visibility=), allow threads: False {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1264.061164] env[62600]: DEBUG nova.virt.hardware [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Flavor limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1264.061330] env[62600]: DEBUG nova.virt.hardware [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Image limits 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1264.061516] env[62600]: DEBUG nova.virt.hardware [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Flavor pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1264.061672] env[62600]: DEBUG nova.virt.hardware [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Image pref 0:0:0 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1264.061826] env[62600]: DEBUG nova.virt.hardware [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62600) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1264.062047] env[62600]: DEBUG nova.virt.hardware [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1264.062218] env[62600]: DEBUG nova.virt.hardware [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1264.062393] env[62600]: DEBUG nova.virt.hardware [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Got 1 possible topologies {{(pid=62600) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1264.062561] env[62600]: DEBUG nova.virt.hardware [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1264.062737] env[62600]: DEBUG nova.virt.hardware [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62600) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1264.063867] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e8c4e9-5511-4bdc-a6fe-f0bd3e1a7e75 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.071663] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b291ca7-2d9f-4af4-aae9-0db6b9099501 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.085123] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:09:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '07e9bef1-2b0e-4e4d-997f-de71bb0e213a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9c4c9eaa-2ecd-4598-bed6-f746cbe53f80', 'vif_model': 'vmxnet3'}] {{(pid=62600) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1264.092262] env[62600]: DEBUG oslo.service.loopingcall [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1264.092470] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Creating VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1264.092660] env[62600]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aecc438c-e88d-41ad-928f-d07db99a83f1 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.111136] env[62600]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1264.111136] env[62600]: value = "task-1223155" [ 1264.111136] env[62600]: _type = "Task" [ 1264.111136] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.118372] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1223155, 'name': CreateVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.620376] env[62600]: DEBUG oslo_vmware.api [-] Task: {'id': task-1223155, 'name': CreateVM_Task, 'duration_secs': 0.285495} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.620537] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Created VM on the ESX host {{(pid=62600) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1264.621203] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/27540ab1-93b4-4ac2-bd01-e836c555bff5" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1264.621372] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquired lock "[datastore2] devstack-image-cache_base/27540ab1-93b4-4ac2-bd01-e836c555bff5" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1264.621742] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/27540ab1-93b4-4ac2-bd01-e836c555bff5" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1264.621992] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e30df91-4c49-40c9-85cd-e8daaa00b04f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.626470] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1264.626470] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520fe427-0efd-0647-62a2-5be67862c2f0" [ 1264.626470] env[62600]: _type = "Task" [ 1264.626470] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.633794] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520fe427-0efd-0647-62a2-5be67862c2f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.722932] env[62600]: DEBUG nova.compute.manager [req-d398dd2d-640f-4a5f-aebf-34025d901bda req-ba59f9c7-551a-4043-a032-e032800b4f79 service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Received event network-changed-9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1264.723148] env[62600]: DEBUG nova.compute.manager [req-d398dd2d-640f-4a5f-aebf-34025d901bda req-ba59f9c7-551a-4043-a032-e032800b4f79 service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Refreshing instance network info cache due to event network-changed-9c4c9eaa-2ecd-4598-bed6-f746cbe53f80. {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1264.723323] env[62600]: DEBUG oslo_concurrency.lockutils [req-d398dd2d-640f-4a5f-aebf-34025d901bda req-ba59f9c7-551a-4043-a032-e032800b4f79 service nova] Acquiring lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1264.723482] env[62600]: DEBUG oslo_concurrency.lockutils [req-d398dd2d-640f-4a5f-aebf-34025d901bda req-ba59f9c7-551a-4043-a032-e032800b4f79 service nova] Acquired lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1264.723684] env[62600]: DEBUG nova.network.neutron [req-d398dd2d-640f-4a5f-aebf-34025d901bda req-ba59f9c7-551a-4043-a032-e032800b4f79 service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Refreshing network info cache for port 9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1265.136497] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Releasing lock "[datastore2] devstack-image-cache_base/27540ab1-93b4-4ac2-bd01-e836c555bff5" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1265.136851] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Processing image 27540ab1-93b4-4ac2-bd01-e836c555bff5 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1265.136978] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/27540ab1-93b4-4ac2-bd01-e836c555bff5/27540ab1-93b4-4ac2-bd01-e836c555bff5.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1265.137156] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquired lock "[datastore2] devstack-image-cache_base/27540ab1-93b4-4ac2-bd01-e836c555bff5/27540ab1-93b4-4ac2-bd01-e836c555bff5.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1265.137345] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1265.137589] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c34bea41-a61c-49e7-908e-b9c73a937e22 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.147316] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1265.147495] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62600) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1265.148205] env[62600]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2217bec8-a5ca-4ad1-a016-92ea979351bc {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.152628] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1265.152628] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520abb4a-cc25-58db-ed01-f5c83e40092c" [ 1265.152628] env[62600]: _type = "Task" [ 1265.152628] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.159591] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': session[523e801c-4d13-2c46-e8f5-3e004088fa0b]520abb4a-cc25-58db-ed01-f5c83e40092c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.417586] env[62600]: DEBUG nova.network.neutron [req-d398dd2d-640f-4a5f-aebf-34025d901bda req-ba59f9c7-551a-4043-a032-e032800b4f79 service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Updated VIF entry in instance network info cache for port 9c4c9eaa-2ecd-4598-bed6-f746cbe53f80. {{(pid=62600) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1265.417958] env[62600]: DEBUG nova.network.neutron [req-d398dd2d-640f-4a5f-aebf-34025d901bda req-ba59f9c7-551a-4043-a032-e032800b4f79 service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Updating instance_info_cache with network_info: [{"id": "9c4c9eaa-2ecd-4598-bed6-f746cbe53f80", "address": "fa:16:3e:6b:09:8c", "network": {"id": "dc21424f-b891-428d-a74e-76ab78d757a8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1428037886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "952323d62f034126a9b46750d03fe6ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c4c9eaa-2e", "ovs_interfaceid": "9c4c9eaa-2ecd-4598-bed6-f746cbe53f80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1265.662033] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Preparing fetch location {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1265.662321] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Fetch image to [datastore2] OSTACK_IMG_6d1ec185-9419-478f-ac84-2a1d92e5b862/OSTACK_IMG_6d1ec185-9419-478f-ac84-2a1d92e5b862.vmdk {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1265.662513] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Downloading stream optimized image 27540ab1-93b4-4ac2-bd01-e836c555bff5 to [datastore2] OSTACK_IMG_6d1ec185-9419-478f-ac84-2a1d92e5b862/OSTACK_IMG_6d1ec185-9419-478f-ac84-2a1d92e5b862.vmdk on the data store datastore2 as vApp {{(pid=62600) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1265.662689] env[62600]: DEBUG nova.virt.vmwareapi.images [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Downloading image file data 27540ab1-93b4-4ac2-bd01-e836c555bff5 to the ESX as VM named 'OSTACK_IMG_6d1ec185-9419-478f-ac84-2a1d92e5b862' {{(pid=62600) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1265.727151] env[62600]: DEBUG oslo_vmware.rw_handles [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1265.727151] env[62600]: value = "resgroup-9" [ 1265.727151] env[62600]: _type = "ResourcePool" [ 1265.727151] env[62600]: }. {{(pid=62600) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1265.727451] env[62600]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-0d93d526-26d2-416c-bcf1-58bad851db2f {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.747929] env[62600]: DEBUG oslo_vmware.rw_handles [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lease: (returnval){ [ 1265.747929] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52fc70fe-6a70-718b-bc75-91e2cc0e6566" [ 1265.747929] env[62600]: _type = "HttpNfcLease" [ 1265.747929] env[62600]: } obtained for vApp import into resource pool (val){ [ 1265.747929] env[62600]: value = "resgroup-9" [ 1265.747929] env[62600]: _type = "ResourcePool" [ 1265.747929] env[62600]: }. {{(pid=62600) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1265.748298] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the lease: (returnval){ [ 1265.748298] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52fc70fe-6a70-718b-bc75-91e2cc0e6566" [ 1265.748298] env[62600]: _type = "HttpNfcLease" [ 1265.748298] env[62600]: } to be ready. {{(pid=62600) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1265.754442] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1265.754442] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52fc70fe-6a70-718b-bc75-91e2cc0e6566" [ 1265.754442] env[62600]: _type = "HttpNfcLease" [ 1265.754442] env[62600]: } is initializing. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1265.920550] env[62600]: DEBUG oslo_concurrency.lockutils [req-d398dd2d-640f-4a5f-aebf-34025d901bda req-ba59f9c7-551a-4043-a032-e032800b4f79 service nova] Releasing lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1266.256551] env[62600]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1266.256551] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52fc70fe-6a70-718b-bc75-91e2cc0e6566" [ 1266.256551] env[62600]: _type = "HttpNfcLease" [ 1266.256551] env[62600]: } is ready. {{(pid=62600) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1266.256986] env[62600]: DEBUG oslo_vmware.rw_handles [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1266.256986] env[62600]: value = "session[523e801c-4d13-2c46-e8f5-3e004088fa0b]52fc70fe-6a70-718b-bc75-91e2cc0e6566" [ 1266.256986] env[62600]: _type = "HttpNfcLease" [ 1266.256986] env[62600]: }. {{(pid=62600) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1266.257559] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e29dfc6-07a6-455c-9ca9-a818532c8223 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.265781] env[62600]: DEBUG oslo_vmware.rw_handles [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527a1df7-1655-7b6a-b1c4-b593717dffd5/disk-0.vmdk from lease info. {{(pid=62600) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1266.265983] env[62600]: DEBUG oslo_vmware.rw_handles [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527a1df7-1655-7b6a-b1c4-b593717dffd5/disk-0.vmdk. {{(pid=62600) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1266.328863] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2a8d6007-2264-411c-b411-fa84aaaabb3b {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.395149] env[62600]: DEBUG oslo_vmware.rw_handles [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Completed reading data from the image iterator. {{(pid=62600) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1267.395149] env[62600]: DEBUG oslo_vmware.rw_handles [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527a1df7-1655-7b6a-b1c4-b593717dffd5/disk-0.vmdk. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1267.395149] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fc306ee-8aea-4f43-af91-7a0b51c571b6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.402371] env[62600]: DEBUG oslo_vmware.rw_handles [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527a1df7-1655-7b6a-b1c4-b593717dffd5/disk-0.vmdk is in state: ready. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1267.402781] env[62600]: DEBUG oslo_vmware.rw_handles [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527a1df7-1655-7b6a-b1c4-b593717dffd5/disk-0.vmdk. {{(pid=62600) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1267.403208] env[62600]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-7cc3c3e9-4df9-4558-b8b7-0cf4d4982b11 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.587981] env[62600]: DEBUG oslo_vmware.rw_handles [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527a1df7-1655-7b6a-b1c4-b593717dffd5/disk-0.vmdk. {{(pid=62600) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1267.587981] env[62600]: INFO nova.virt.vmwareapi.images [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Downloaded image file data 27540ab1-93b4-4ac2-bd01-e836c555bff5 [ 1267.588329] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8cd70f-faec-4471-b1e4-b0859afd7348 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.602950] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d3683f4f-557d-40a8-862c-b6d556ab2497 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.639424] env[62600]: INFO nova.virt.vmwareapi.images [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] The imported VM was unregistered [ 1267.641970] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Caching image {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1267.642259] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Creating directory with path [datastore2] devstack-image-cache_base/27540ab1-93b4-4ac2-bd01-e836c555bff5 {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1267.642559] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4ae33fc8-f5e3-442f-9031-39b79f22b0e2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.661651] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Created directory with path [datastore2] devstack-image-cache_base/27540ab1-93b4-4ac2-bd01-e836c555bff5 {{(pid=62600) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1267.661845] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_6d1ec185-9419-478f-ac84-2a1d92e5b862/OSTACK_IMG_6d1ec185-9419-478f-ac84-2a1d92e5b862.vmdk to [datastore2] devstack-image-cache_base/27540ab1-93b4-4ac2-bd01-e836c555bff5/27540ab1-93b4-4ac2-bd01-e836c555bff5.vmdk. {{(pid=62600) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1267.662127] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-f79f8877-5347-45a0-a544-c713932ee95d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.668353] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1267.668353] env[62600]: value = "task-1223158" [ 1267.668353] env[62600]: _type = "Task" [ 1267.668353] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.675415] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223158, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.178154] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223158, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.678901] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223158, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.180231] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223158, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.681214] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223158, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.182010] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223158, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.222956} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.182325] env[62600]: INFO nova.virt.vmwareapi.ds_util [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_6d1ec185-9419-478f-ac84-2a1d92e5b862/OSTACK_IMG_6d1ec185-9419-478f-ac84-2a1d92e5b862.vmdk to [datastore2] devstack-image-cache_base/27540ab1-93b4-4ac2-bd01-e836c555bff5/27540ab1-93b4-4ac2-bd01-e836c555bff5.vmdk. [ 1270.182519] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Cleaning up location [datastore2] OSTACK_IMG_6d1ec185-9419-478f-ac84-2a1d92e5b862 {{(pid=62600) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1270.182727] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_6d1ec185-9419-478f-ac84-2a1d92e5b862 {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1270.182978] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c45a6924-9a78-4a04-8e3e-8e9afc9248ce {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.188955] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1270.188955] env[62600]: value = "task-1223159" [ 1270.188955] env[62600]: _type = "Task" [ 1270.188955] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.196110] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223159, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.698884] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223159, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.082207} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.699282] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1270.699329] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Releasing lock "[datastore2] devstack-image-cache_base/27540ab1-93b4-4ac2-bd01-e836c555bff5/27540ab1-93b4-4ac2-bd01-e836c555bff5.vmdk" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1270.699587] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/27540ab1-93b4-4ac2-bd01-e836c555bff5/27540ab1-93b4-4ac2-bd01-e836c555bff5.vmdk to [datastore2] 00648d57-68ae-438b-af37-7e50d6cdcedd/00648d57-68ae-438b-af37-7e50d6cdcedd.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1270.699830] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ab3227b6-d5b8-4f91-b168-d662437ec2a4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.706956] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1270.706956] env[62600]: value = "task-1223160" [ 1270.706956] env[62600]: _type = "Task" [ 1270.706956] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.713983] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223160, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.216706] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223160, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.718841] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223160, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.218257] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223160, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.720283] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223160, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1273.219543] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223160, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.17807} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1273.219831] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/27540ab1-93b4-4ac2-bd01-e836c555bff5/27540ab1-93b4-4ac2-bd01-e836c555bff5.vmdk to [datastore2] 00648d57-68ae-438b-af37-7e50d6cdcedd/00648d57-68ae-438b-af37-7e50d6cdcedd.vmdk {{(pid=62600) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1273.220585] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69345f9d-082c-4466-8ef6-57f15411c51e {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.241403] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 00648d57-68ae-438b-af37-7e50d6cdcedd/00648d57-68ae-438b-af37-7e50d6cdcedd.vmdk or device None with type streamOptimized {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1273.241640] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35fe673f-8fd9-45d0-a3fc-7a2d2937ac40 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.259897] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1273.259897] env[62600]: value = "task-1223161" [ 1273.259897] env[62600]: _type = "Task" [ 1273.259897] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1273.266875] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223161, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1273.769703] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223161, 'name': ReconfigVM_Task, 'duration_secs': 0.287115} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1273.770126] env[62600]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 00648d57-68ae-438b-af37-7e50d6cdcedd/00648d57-68ae-438b-af37-7e50d6cdcedd.vmdk or device None with type streamOptimized {{(pid=62600) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1273.770569] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c089bdcb-fcb7-4b56-865d-2ea1498e6c10 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.776505] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1273.776505] env[62600]: value = "task-1223162" [ 1273.776505] env[62600]: _type = "Task" [ 1273.776505] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1273.783548] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223162, 'name': Rename_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.286090] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223162, 'name': Rename_Task, 'duration_secs': 0.167719} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.286380] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Powering on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1274.286637] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-19402818-1d61-466e-8819-a0f7e57f8ab4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.292988] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1274.292988] env[62600]: value = "task-1223163" [ 1274.292988] env[62600]: _type = "Task" [ 1274.292988] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.299907] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223163, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.803192] env[62600]: DEBUG oslo_vmware.api [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223163, 'name': PowerOnVM_Task, 'duration_secs': 0.45141} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.803595] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Powered on the VM {{(pid=62600) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1274.892929] env[62600]: DEBUG nova.compute.manager [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Checking state {{(pid=62600) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1274.893894] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae46cb3b-8305-4263-bf84-7e7f05412ca8 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.409661] env[62600]: DEBUG oslo_concurrency.lockutils [None req-2ce3a076-571f-4178-a287-78fe2876893f tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "00648d57-68ae-438b-af37-7e50d6cdcedd" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.252s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1278.728651] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1280.727695] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1281.727697] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1281.728107] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1281.728107] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1281.728236] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1282.231687] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1282.231946] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1282.232096] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1282.232255] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62600) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1282.233194] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e25c9a2-1a93-466c-af6c-8ce0f4d3ca59 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.241171] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-580114e7-e257-47d4-a3a2-445192f8898d {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.255821] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5736d573-1d68-428a-bcf0-06fe08c3b027 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.261849] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fda8789-98fa-4380-858d-e58a17e32e90 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.289240] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181032MB free_disk=152GB free_vcpus=48 pci_devices=None {{(pid=62600) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1282.289405] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1282.289586] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1283.312917] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Instance 00648d57-68ae-438b-af37-7e50d6cdcedd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62600) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1283.313180] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1283.313281] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62600) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1283.336685] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfb408c7-b917-4e9f-bd63-f43e8ac893d6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.343978] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5c2258a-f8c1-4430-9e13-ca23d63dfc51 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.373506] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fcf3299-4961-4840-a93b-dfa88d0d9eb2 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.380248] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e303b161-e144-4203-a77c-5e84dcdb13ff {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.392989] env[62600]: DEBUG nova.compute.provider_tree [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1283.896166] env[62600]: DEBUG nova.scheduler.client.report [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1284.400976] env[62600]: DEBUG nova.compute.resource_tracker [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62600) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1284.401343] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.112s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1286.397435] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1286.397843] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1286.397843] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Starting heal instance info cache {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1286.397963] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Rebuilding the list of instances to heal {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1286.932450] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquiring lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1286.932596] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Acquired lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1286.932740] env[62600]: DEBUG nova.network.neutron [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Forcefully refreshing network info cache for instance {{(pid=62600) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1286.932930] env[62600]: DEBUG nova.objects.instance [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Lazy-loading 'info_cache' on Instance uuid 00648d57-68ae-438b-af37-7e50d6cdcedd {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1288.646980] env[62600]: DEBUG nova.network.neutron [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Updating instance_info_cache with network_info: [{"id": "9c4c9eaa-2ecd-4598-bed6-f746cbe53f80", "address": "fa:16:3e:6b:09:8c", "network": {"id": "dc21424f-b891-428d-a74e-76ab78d757a8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1428037886-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "952323d62f034126a9b46750d03fe6ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "07e9bef1-2b0e-4e4d-997f-de71bb0e213a", "external-id": "nsx-vlan-transportzone-786", "segmentation_id": 786, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c4c9eaa-2e", "ovs_interfaceid": "9c4c9eaa-2ecd-4598-bed6-f746cbe53f80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1289.150143] env[62600]: DEBUG oslo_concurrency.lockutils [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Releasing lock "refresh_cache-00648d57-68ae-438b-af37-7e50d6cdcedd" {{(pid=62600) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1289.150391] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Updated the network info_cache for instance {{(pid=62600) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1289.150586] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1289.150723] env[62600]: DEBUG nova.compute.manager [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62600) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1312.297215] env[62600]: DEBUG oslo_concurrency.lockutils [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "00648d57-68ae-438b-af37-7e50d6cdcedd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1312.297604] env[62600]: DEBUG oslo_concurrency.lockutils [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "00648d57-68ae-438b-af37-7e50d6cdcedd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1312.297777] env[62600]: DEBUG oslo_concurrency.lockutils [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "00648d57-68ae-438b-af37-7e50d6cdcedd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1312.298024] env[62600]: DEBUG oslo_concurrency.lockutils [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "00648d57-68ae-438b-af37-7e50d6cdcedd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1312.298254] env[62600]: DEBUG oslo_concurrency.lockutils [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "00648d57-68ae-438b-af37-7e50d6cdcedd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1312.301698] env[62600]: INFO nova.compute.manager [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Terminating instance [ 1312.303630] env[62600]: DEBUG nova.compute.manager [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Start destroying the instance on the hypervisor. {{(pid=62600) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1312.303869] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Destroying instance {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1312.304858] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfd52aca-35a7-4f36-9c4b-3982ac8e7cf7 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.313942] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Powering off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1312.314236] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-85eacdc3-ef5e-41b4-9867-24af76c3b52a {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.320313] env[62600]: DEBUG oslo_vmware.api [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1312.320313] env[62600]: value = "task-1223164" [ 1312.320313] env[62600]: _type = "Task" [ 1312.320313] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1312.327655] env[62600]: DEBUG oslo_vmware.api [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223164, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1312.830504] env[62600]: DEBUG oslo_vmware.api [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223164, 'name': PowerOffVM_Task, 'duration_secs': 0.193599} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1312.830794] env[62600]: DEBUG nova.virt.vmwareapi.vm_util [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Powered off the VM {{(pid=62600) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1312.830972] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Unregistering the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1312.831260] env[62600]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d323d3bf-bc40-41cc-b8e3-2ac6a05131e6 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.893044] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Unregistered the VM {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1312.893324] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Deleting contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1312.893483] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Deleting the datastore file [datastore2] 00648d57-68ae-438b-af37-7e50d6cdcedd {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1312.893758] env[62600]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bd45be94-8d56-4563-8020-6b8c30073560 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.899892] env[62600]: DEBUG oslo_vmware.api [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for the task: (returnval){ [ 1312.899892] env[62600]: value = "task-1223166" [ 1312.899892] env[62600]: _type = "Task" [ 1312.899892] env[62600]: } to complete. {{(pid=62600) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1312.906787] env[62600]: DEBUG oslo_vmware.api [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223166, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1313.410471] env[62600]: DEBUG oslo_vmware.api [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Task: {'id': task-1223166, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.119637} completed successfully. {{(pid=62600) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1313.410907] env[62600]: DEBUG nova.virt.vmwareapi.ds_util [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Deleted the datastore file {{(pid=62600) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1313.410907] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Deleted contents of the VM from datastore datastore2 {{(pid=62600) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1313.411107] env[62600]: DEBUG nova.virt.vmwareapi.vmops [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Instance destroyed {{(pid=62600) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1313.411258] env[62600]: INFO nova.compute.manager [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1313.411519] env[62600]: DEBUG oslo.service.loopingcall [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62600) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1313.411707] env[62600]: DEBUG nova.compute.manager [-] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Deallocating network for instance {{(pid=62600) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1313.411800] env[62600]: DEBUG nova.network.neutron [-] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] deallocate_for_instance() {{(pid=62600) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1313.813204] env[62600]: DEBUG nova.compute.manager [req-7f56010c-4f3a-4ef4-8dc9-a7b3adb67ab0 req-ec461e87-91c2-4070-b978-c91300eb32dc service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Received event network-vif-deleted-9c4c9eaa-2ecd-4598-bed6-f746cbe53f80 {{(pid=62600) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1313.813274] env[62600]: INFO nova.compute.manager [req-7f56010c-4f3a-4ef4-8dc9-a7b3adb67ab0 req-ec461e87-91c2-4070-b978-c91300eb32dc service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Neutron deleted interface 9c4c9eaa-2ecd-4598-bed6-f746cbe53f80; detaching it from the instance and deleting it from the info cache [ 1313.813436] env[62600]: DEBUG nova.network.neutron [req-7f56010c-4f3a-4ef4-8dc9-a7b3adb67ab0 req-ec461e87-91c2-4070-b978-c91300eb32dc service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1314.290074] env[62600]: DEBUG nova.network.neutron [-] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Updating instance_info_cache with network_info: [] {{(pid=62600) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1314.315849] env[62600]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1ff73945-d56e-46af-ad2b-332e139d53b4 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.324955] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58efa000-9fb6-41ba-a7cd-1ea096873b6c {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.347573] env[62600]: DEBUG nova.compute.manager [req-7f56010c-4f3a-4ef4-8dc9-a7b3adb67ab0 req-ec461e87-91c2-4070-b978-c91300eb32dc service nova] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Detach interface failed, port_id=9c4c9eaa-2ecd-4598-bed6-f746cbe53f80, reason: Instance 00648d57-68ae-438b-af37-7e50d6cdcedd could not be found. {{(pid=62600) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1314.793392] env[62600]: INFO nova.compute.manager [-] [instance: 00648d57-68ae-438b-af37-7e50d6cdcedd] Took 1.38 seconds to deallocate network for instance. [ 1315.299752] env[62600]: DEBUG oslo_concurrency.lockutils [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1315.300052] env[62600]: DEBUG oslo_concurrency.lockutils [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1315.300305] env[62600]: DEBUG nova.objects.instance [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lazy-loading 'resources' on Instance uuid 00648d57-68ae-438b-af37-7e50d6cdcedd {{(pid=62600) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1315.835184] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a900f2d-309f-432a-a1a5-f8798940dc88 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.843938] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4101fac-ab77-4462-8b02-e63f6b6d86e5 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.872535] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c1c6782-a74b-4641-9891-295a32131b64 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.879707] env[62600]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf2087a-2cd5-49ca-97b1-07e019cada19 {{(pid=62600) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.892336] env[62600]: DEBUG nova.compute.provider_tree [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Inventory has not changed in ProviderTree for provider: 664af347-7147-4bf5-9019-9ae15cb4aa82 {{(pid=62600) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1316.395111] env[62600]: DEBUG nova.scheduler.client.report [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Inventory has not changed for provider 664af347-7147-4bf5-9019-9ae15cb4aa82 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 152, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62600) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1316.900898] env[62600]: DEBUG oslo_concurrency.lockutils [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.601s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1316.920767] env[62600]: INFO nova.scheduler.client.report [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Deleted allocations for instance 00648d57-68ae-438b-af37-7e50d6cdcedd [ 1317.428025] env[62600]: DEBUG oslo_concurrency.lockutils [None req-387fd64f-bc8f-44cd-a964-f0ec170e1d03 tempest-AttachVolumeShelveTestJSON-2046149534 tempest-AttachVolumeShelveTestJSON-2046149534-project-member] Lock "00648d57-68ae-438b-af37-7e50d6cdcedd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.130s {{(pid=62600) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1340.727931] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1341.729024] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1341.729024] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1342.727814] env[62600]: DEBUG oslo_service.periodic_task [None req-5d341861-a52d-41ad-a301-da9668e6ff8d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62600) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}}